var/home/core/zuul-output/0000755000175000017500000000000015111054345014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111067334015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005235573715111067325017717 0ustar rootrootNov 24 13:14:17 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 13:14:17 crc restorecon[4584]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:17 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:14:18 crc restorecon[4584]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 13:14:18 crc kubenswrapper[4824]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:14:18 crc kubenswrapper[4824]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 13:14:18 crc kubenswrapper[4824]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:14:18 crc kubenswrapper[4824]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:14:18 crc kubenswrapper[4824]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 13:14:18 crc kubenswrapper[4824]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.788411 4824 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798480 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798551 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798571 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798594 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798612 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798629 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798644 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798658 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798672 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798684 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798698 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798708 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798719 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798729 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798742 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798756 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798766 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798777 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798787 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798796 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798838 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798847 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798857 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798867 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798877 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798897 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798906 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798915 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798923 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798931 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798939 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798947 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798954 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798983 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798991 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.798998 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799007 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799014 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799022 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799029 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799037 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799045 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799053 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799061 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799068 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799076 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799084 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799092 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799099 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799108 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799118 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799126 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799134 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799142 4824 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799149 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799158 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799166 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799175 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799183 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799191 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799199 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799206 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799214 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799221 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799230 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799238 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799245 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799256 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799263 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799271 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.799279 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799467 4824 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799487 4824 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799509 4824 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799521 4824 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799533 4824 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799543 4824 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799555 4824 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799572 4824 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799582 4824 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799591 4824 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799601 4824 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799624 4824 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799634 4824 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799643 4824 flags.go:64] FLAG: --cgroup-root="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799651 4824 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799663 4824 flags.go:64] FLAG: --client-ca-file="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799674 4824 flags.go:64] FLAG: --cloud-config="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799686 4824 flags.go:64] FLAG: --cloud-provider="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799697 4824 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799729 4824 flags.go:64] FLAG: --cluster-domain="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799747 4824 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799759 4824 flags.go:64] FLAG: --config-dir="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799770 4824 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799783 4824 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799799 4824 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799847 4824 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799860 4824 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799872 4824 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799884 4824 flags.go:64] FLAG: --contention-profiling="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799895 4824 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799907 4824 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799919 4824 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799928 4824 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799941 4824 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799950 4824 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799959 4824 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799968 4824 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799977 4824 flags.go:64] FLAG: --enable-server="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.799986 4824 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800010 4824 flags.go:64] FLAG: --event-burst="100" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800020 4824 flags.go:64] FLAG: --event-qps="50" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800029 4824 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800038 4824 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800047 4824 flags.go:64] FLAG: --eviction-hard="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800059 4824 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800068 4824 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800077 4824 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800101 4824 flags.go:64] FLAG: --eviction-soft="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800111 4824 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800121 4824 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800130 4824 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800139 4824 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800147 4824 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800156 4824 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800165 4824 flags.go:64] FLAG: --feature-gates="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800176 4824 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800185 4824 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800195 4824 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800205 4824 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800214 4824 flags.go:64] FLAG: --healthz-port="10248" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800223 4824 flags.go:64] FLAG: --help="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800232 4824 flags.go:64] FLAG: --hostname-override="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800241 4824 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800250 4824 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800260 4824 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800272 4824 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800288 4824 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800310 4824 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800323 4824 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800334 4824 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800347 4824 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800360 4824 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800371 4824 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800380 4824 flags.go:64] FLAG: --kube-reserved="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800389 4824 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800398 4824 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800407 4824 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800419 4824 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800428 4824 flags.go:64] FLAG: --lock-file="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800437 4824 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800447 4824 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800456 4824 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800483 4824 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800504 4824 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800513 4824 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800522 4824 flags.go:64] FLAG: --logging-format="text" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800531 4824 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800541 4824 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800550 4824 flags.go:64] FLAG: --manifest-url="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800604 4824 flags.go:64] FLAG: --manifest-url-header="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800619 4824 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800629 4824 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800640 4824 flags.go:64] FLAG: --max-pods="110" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800650 4824 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800660 4824 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800670 4824 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800680 4824 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800690 4824 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800699 4824 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800708 4824 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800731 4824 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800741 4824 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800750 4824 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800761 4824 flags.go:64] FLAG: --pod-cidr="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800770 4824 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800785 4824 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800796 4824 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800838 4824 flags.go:64] FLAG: --pods-per-core="0" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800850 4824 flags.go:64] FLAG: --port="10250" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800862 4824 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800873 4824 flags.go:64] FLAG: --provider-id="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800885 4824 flags.go:64] FLAG: --qos-reserved="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800896 4824 flags.go:64] FLAG: --read-only-port="10255" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800905 4824 flags.go:64] FLAG: --register-node="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800914 4824 flags.go:64] FLAG: --register-schedulable="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800923 4824 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800939 4824 flags.go:64] FLAG: --registry-burst="10" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800948 4824 flags.go:64] FLAG: --registry-qps="5" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800957 4824 flags.go:64] FLAG: --reserved-cpus="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800980 4824 flags.go:64] FLAG: --reserved-memory="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.800993 4824 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801003 4824 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801012 4824 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801021 4824 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801029 4824 flags.go:64] FLAG: --runonce="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801038 4824 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801048 4824 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801057 4824 flags.go:64] FLAG: --seccomp-default="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801067 4824 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801075 4824 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801085 4824 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801094 4824 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801104 4824 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801113 4824 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801122 4824 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801132 4824 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801142 4824 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801152 4824 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801164 4824 flags.go:64] FLAG: --system-cgroups="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801183 4824 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801211 4824 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801223 4824 flags.go:64] FLAG: --tls-cert-file="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801236 4824 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801272 4824 flags.go:64] FLAG: --tls-min-version="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801283 4824 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801294 4824 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801305 4824 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801315 4824 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801324 4824 flags.go:64] FLAG: --v="2" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801336 4824 flags.go:64] FLAG: --version="false" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801348 4824 flags.go:64] FLAG: --vmodule="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801358 4824 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.801367 4824 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801602 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801612 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801631 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801641 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801649 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801659 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801667 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801680 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801691 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801699 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801709 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801717 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801726 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801734 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801744 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801753 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801762 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801771 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801779 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801787 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801796 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801835 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801844 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801853 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801863 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801871 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801879 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801887 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801894 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801902 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801911 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801921 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801941 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801955 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801965 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801975 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801986 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.801996 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802030 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802045 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802058 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802072 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802084 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802094 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802104 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802113 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802126 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802140 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802151 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802162 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802172 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802182 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802191 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802198 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802206 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802214 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802222 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802230 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802238 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802245 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802253 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802261 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802270 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802278 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802286 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802293 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802301 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802309 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802317 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802324 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.802332 4824 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.802346 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.813332 4824 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.813365 4824 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813436 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813444 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813449 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813453 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813458 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813463 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813469 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813473 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813478 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813482 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813487 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813491 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813495 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813499 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813503 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813507 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813510 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813514 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813518 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813522 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813525 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813529 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813532 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813536 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813540 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813543 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813547 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813550 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813554 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813558 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813563 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813568 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813573 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813577 4824 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813582 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813586 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813590 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813594 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813598 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813602 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813606 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813610 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813613 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813617 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813620 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813624 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813628 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813631 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813635 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813639 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813642 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813646 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813650 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813653 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813658 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813662 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813667 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813671 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813675 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813678 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813682 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813685 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813689 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813692 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813696 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813700 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813703 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813707 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813710 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813714 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813722 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.813729 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813882 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813891 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813896 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813899 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813903 4824 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813907 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813911 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813914 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813918 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813921 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813924 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813928 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813931 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813935 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813938 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813943 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813949 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813953 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813957 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813960 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813964 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813968 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813972 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813975 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813980 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813986 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813990 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813994 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.813998 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814002 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814005 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814010 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814014 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814018 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814023 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814027 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814031 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814035 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814038 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814041 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814045 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814048 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814052 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814056 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814059 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814062 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814066 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814069 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814074 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814077 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814080 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814084 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814087 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814091 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814095 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814099 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814102 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814106 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814109 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814113 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814116 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814119 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814123 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814127 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814130 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814135 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814139 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814143 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814148 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814151 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.814155 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.814161 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.814320 4824 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.819903 4824 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.820026 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.822067 4824 server.go:997] "Starting client certificate rotation" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.822118 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.822252 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-24 07:50:44.727783457 +0000 UTC Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.822323 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.849134 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.852858 4824 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 13:14:18 crc kubenswrapper[4824]: E1124 13:14:18.853093 4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.866948 4824 log.go:25] "Validated CRI v1 runtime API" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.901568 4824 log.go:25] "Validated CRI v1 image API" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.903689 4824 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.908024 4824 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-13-07-34-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.908109 4824 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.918984 4824 manager.go:217] Machine: {Timestamp:2025-11-24 13:14:18.916855898 +0000 UTC m=+0.556395228 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199472640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8f1e113f-2cfe-4be7-935e-94a6d127c173 BootID:d50fc599-297b-4acb-83a1-35b8955262d5 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:3076107 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599734272 Type:vfs Inodes:3076107 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039894528 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:fe:04:fc Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:fe:04:fc Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:63:7f:0f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ac:29:e9 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:50:38:bf Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ef:57:8d Speed:-1 Mtu:1496} {Name:eth10 MacAddress:02:d7:0c:6a:45:a9 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:86:1f:af:39:52:e9 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199472640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.919287 4824 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.919473 4824 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.920719 4824 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.920943 4824 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.921081 4824 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.921359 4824 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.921416 4824 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.921978 4824 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.922062 4824 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.922269 4824 state_mem.go:36] "Initialized new in-memory state store" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.922406 4824 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.925968 4824 kubelet.go:418] "Attempting to sync node with API server" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.926041 4824 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.926127 4824 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.926208 4824 kubelet.go:324] "Adding apiserver pod source" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.926301 4824 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.930064 4824 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.930977 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.932326 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.932351 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:18 crc kubenswrapper[4824]: E1124 13:14:18.932545 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:18 crc kubenswrapper[4824]: E1124 13:14:18.932426 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.933664 4824 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935324 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935411 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935478 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935531 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935584 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935649 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935710 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935765 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935845 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935904 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.935999 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.936060 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.937115 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.937602 4824 server.go:1280] "Started kubelet" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.938476 4824 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.938994 4824 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.938481 4824 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.939337 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:18 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.939733 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.939757 4824 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.940007 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 07:25:52.678240492 +0000 UTC Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.940107 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 762h11m33.738135994s for next certificate rotation Nov 24 13:14:18 crc kubenswrapper[4824]: E1124 13:14:18.940447 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.940730 4824 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.940744 4824 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.940940 4824 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.941223 4824 factory.go:55] Registering systemd factory Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.941269 4824 factory.go:221] Registration of the systemd container factory successfully Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.942176 4824 factory.go:153] Registering CRI-O factory Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.942196 4824 factory.go:221] Registration of the crio container factory successfully Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.942293 4824 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.945239 4824 server.go:460] "Adding debug handlers to kubelet server" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.944885 4824 factory.go:103] Registering Raw factory Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.946218 4824 manager.go:1196] Started watching for new ooms in manager Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.948916 4824 manager.go:319] Starting recovery of all containers Nov 24 13:14:18 crc kubenswrapper[4824]: E1124 13:14:18.949451 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.188:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187af39cd24b6598 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 13:14:18.937574808 +0000 UTC m=+0.577114118,LastTimestamp:2025-11-24 13:14:18.937574808 +0000 UTC m=+0.577114118,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 13:14:18 crc kubenswrapper[4824]: W1124 13:14:18.952890 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:18 crc kubenswrapper[4824]: E1124 13:14:18.953275 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:18 crc kubenswrapper[4824]: E1124 13:14:18.959084 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.188:6443: connect: connection refused" interval="200ms" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.959999 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960064 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960079 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960092 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960106 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960120 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960134 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960148 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960164 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960206 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960220 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960234 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960257 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960302 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960349 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960366 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960407 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960428 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960446 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960485 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960503 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960521 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960538 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960579 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960599 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960617 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960728 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960753 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960774 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960792 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960857 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960885 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960904 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960931 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960950 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.960990 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961008 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961028 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961046 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961072 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961094 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961112 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961130 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961149 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961169 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961188 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961208 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961224 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961241 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961261 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961278 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961295 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961319 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961340 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961368 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961390 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961417 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961436 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961453 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961473 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961497 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961521 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961538 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961557 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961575 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961592 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961609 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961632 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961650 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961668 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961686 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961705 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961723 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961741 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961760 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961781 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961825 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961845 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961863 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.961883 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.967652 4824 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.967912 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.968050 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.968227 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.968430 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.968591 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.968798 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969523 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969587 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969614 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969635 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969651 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969671 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969689 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969703 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969723 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969738 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969753 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969770 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969799 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969832 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969846 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969864 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969885 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969902 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969930 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969954 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.969998 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970035 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970061 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970085 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970101 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970116 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970134 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970149 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970163 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970187 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970208 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970253 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970264 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970277 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970289 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970303 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970323 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970337 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970352 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970366 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970383 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970399 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970420 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970437 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970450 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970461 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970473 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970486 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970498 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970514 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970530 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970549 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970566 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970581 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970601 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970616 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970701 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970717 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970729 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970740 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970753 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970765 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970781 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970796 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970827 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.970840 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971539 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971570 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971583 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971594 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971606 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971654 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971669 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971680 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971693 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971704 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971714 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971725 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971740 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971752 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971762 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971772 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971781 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971792 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971892 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971907 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971917 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971929 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971939 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971950 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971962 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971972 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971984 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.971995 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972008 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972018 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972029 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972039 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972050 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972060 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972071 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972082 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972091 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972102 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972114 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972124 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972134 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972143 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972159 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972172 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972193 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972209 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972223 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972244 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972263 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972282 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972297 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972313 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972326 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972338 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972359 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972374 4824 reconstruct.go:97] "Volume reconstruction finished" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.972383 4824 reconciler.go:26] "Reconciler: start to sync state" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.974779 4824 manager.go:324] Recovery completed Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.988540 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.991585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.991706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.991857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.992861 4824 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.992948 4824 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 13:14:18 crc kubenswrapper[4824]: I1124 13:14:18.993038 4824 state_mem.go:36] "Initialized new in-memory state store" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.004832 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.008660 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.008755 4824 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.008796 4824 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.009134 4824 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 13:14:19 crc kubenswrapper[4824]: W1124 13:14:19.009862 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.009919 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.013419 4824 policy_none.go:49] "None policy: Start" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.014776 4824 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.015073 4824 state_mem.go:35] "Initializing new in-memory state store" Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.041223 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.076897 4824 manager.go:334] "Starting Device Plugin manager" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.076986 4824 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.077001 4824 server.go:79] "Starting device plugin registration server" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.077361 4824 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.077379 4824 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.078106 4824 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.078236 4824 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.078243 4824 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.086929 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.109572 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.109656 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.110907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.110981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.110999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.111265 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.112416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.112448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.112462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.112649 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.112715 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.112718 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.113359 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.113407 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.113844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.113885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.113907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.114083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.114119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.114128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.114686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.114706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.114713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.115053 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.115213 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.115245 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.115969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.115972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.116094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.116119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.116061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.116263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.116267 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.116694 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.116732 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.117831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.117864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.117877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.118061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.118085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.118097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.118065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.118223 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.119005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.119036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.119049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.160704 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.188:6443: connect: connection refused" interval="400ms" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175254 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175312 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175331 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175345 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175364 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175527 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175610 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175635 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.175690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.178146 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.179151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.179185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.179195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.179222 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.179687 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.188:6443: connect: connection refused" node="crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277038 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277255 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277305 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277368 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277629 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277436 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277786 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.277983 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278012 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278049 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278067 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278073 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278084 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278086 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278139 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278146 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278155 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.278196 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.380012 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.387469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.387539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.387554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.387588 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.388522 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.188:6443: connect: connection refused" node="crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.439131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.447155 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.466664 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: W1124 13:14:19.484512 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-6587f6bf2186daee54ccc2e178eca1fa03096eb7a6e06296f9c543cf5233d5cf WatchSource:0}: Error finding container 6587f6bf2186daee54ccc2e178eca1fa03096eb7a6e06296f9c543cf5233d5cf: Status 404 returned error can't find the container with id 6587f6bf2186daee54ccc2e178eca1fa03096eb7a6e06296f9c543cf5233d5cf Nov 24 13:14:19 crc kubenswrapper[4824]: W1124 13:14:19.486125 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-de950c2d6913e650fe26b78f82de3b4d8895e49c09d2687cab8cdd7fd8a33444 WatchSource:0}: Error finding container de950c2d6913e650fe26b78f82de3b4d8895e49c09d2687cab8cdd7fd8a33444: Status 404 returned error can't find the container with id de950c2d6913e650fe26b78f82de3b4d8895e49c09d2687cab8cdd7fd8a33444 Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.486500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.493141 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:14:19 crc kubenswrapper[4824]: W1124 13:14:19.495580 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-5544711184b9e6e6d9f6bb3c810c000842cd79286edab81901fecbb68e5a3600 WatchSource:0}: Error finding container 5544711184b9e6e6d9f6bb3c810c000842cd79286edab81901fecbb68e5a3600: Status 404 returned error can't find the container with id 5544711184b9e6e6d9f6bb3c810c000842cd79286edab81901fecbb68e5a3600 Nov 24 13:14:19 crc kubenswrapper[4824]: W1124 13:14:19.503558 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-dd3cee6cfeba537592d38c5fc74419c122e6cd4e7e0c0a3b615277abc8abbf3c WatchSource:0}: Error finding container dd3cee6cfeba537592d38c5fc74419c122e6cd4e7e0c0a3b615277abc8abbf3c: Status 404 returned error can't find the container with id dd3cee6cfeba537592d38c5fc74419c122e6cd4e7e0c0a3b615277abc8abbf3c Nov 24 13:14:19 crc kubenswrapper[4824]: W1124 13:14:19.513450 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-2f159155d9564a605c28c0eaeb7d0f017d9110b469ef4a65ace0ae29c95c6b94 WatchSource:0}: Error finding container 2f159155d9564a605c28c0eaeb7d0f017d9110b469ef4a65ace0ae29c95c6b94: Status 404 returned error can't find the container with id 2f159155d9564a605c28c0eaeb7d0f017d9110b469ef4a65ace0ae29c95c6b94 Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.562415 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.188:6443: connect: connection refused" interval="800ms" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.788912 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.790448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.790491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.790504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.790529 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.791009 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.188:6443: connect: connection refused" node="crc" Nov 24 13:14:19 crc kubenswrapper[4824]: W1124 13:14:19.907252 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:19 crc kubenswrapper[4824]: E1124 13:14:19.907391 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:19 crc kubenswrapper[4824]: I1124 13:14:19.941162 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.014219 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6587f6bf2186daee54ccc2e178eca1fa03096eb7a6e06296f9c543cf5233d5cf"} Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.015563 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2f159155d9564a605c28c0eaeb7d0f017d9110b469ef4a65ace0ae29c95c6b94"} Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.016498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dd3cee6cfeba537592d38c5fc74419c122e6cd4e7e0c0a3b615277abc8abbf3c"} Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.020322 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5544711184b9e6e6d9f6bb3c810c000842cd79286edab81901fecbb68e5a3600"} Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.021873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"de950c2d6913e650fe26b78f82de3b4d8895e49c09d2687cab8cdd7fd8a33444"} Nov 24 13:14:20 crc kubenswrapper[4824]: W1124 13:14:20.349866 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:20 crc kubenswrapper[4824]: E1124 13:14:20.350277 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:20 crc kubenswrapper[4824]: W1124 13:14:20.351417 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:20 crc kubenswrapper[4824]: E1124 13:14:20.351461 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:20 crc kubenswrapper[4824]: E1124 13:14:20.363451 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.188:6443: connect: connection refused" interval="1.6s" Nov 24 13:14:20 crc kubenswrapper[4824]: W1124 13:14:20.428181 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:20 crc kubenswrapper[4824]: E1124 13:14:20.428288 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.591910 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.593456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.593508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.593532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.593587 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:14:20 crc kubenswrapper[4824]: E1124 13:14:20.595895 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.188:6443: connect: connection refused" node="crc" Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.908014 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 13:14:20 crc kubenswrapper[4824]: E1124 13:14:20.909372 4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:20 crc kubenswrapper[4824]: I1124 13:14:20.941126 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.027569 4824 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711" exitCode=0 Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.027660 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711"} Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.027792 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.029240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.029287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.029297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.029663 4824 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c" exitCode=0 Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.029742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c"} Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.029842 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.031569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.031603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.031614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.034082 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06"} Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.034127 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027"} Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.034146 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8"} Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.034162 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c"} Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.034133 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.034926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.034982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.035006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.036616 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd" exitCode=0 Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.036693 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.036681 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd"} Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.037699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.037772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.037795 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.038616 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2" exitCode=0 Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.038653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2"} Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.038768 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.039615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.039656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.039682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.042699 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.043900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.043949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.043966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:21 crc kubenswrapper[4824]: W1124 13:14:21.818755 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:21 crc kubenswrapper[4824]: E1124 13:14:21.818866 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:21 crc kubenswrapper[4824]: I1124 13:14:21.941195 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:21 crc kubenswrapper[4824]: E1124 13:14:21.965020 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.188:6443: connect: connection refused" interval="3.2s" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.045721 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c" exitCode=0 Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.045835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c"} Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.049147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072"} Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.051552 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3"} Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.053400 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.053705 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c"} Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.055867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.055915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.055931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:22 crc kubenswrapper[4824]: W1124 13:14:22.178918 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:22 crc kubenswrapper[4824]: E1124 13:14:22.179034 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.196401 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.209253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.209315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.209324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.209354 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:14:22 crc kubenswrapper[4824]: E1124 13:14:22.209997 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.188:6443: connect: connection refused" node="crc" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.508603 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:22 crc kubenswrapper[4824]: I1124 13:14:22.940469 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.060368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c"} Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.060441 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c"} Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.060437 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.061507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.061540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.061551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.063486 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f"} Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.063517 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b"} Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.063528 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a"} Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.063556 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.063580 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.064196 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.064839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.064868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.064880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.064906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.064937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.064954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.065792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.065839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.065852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:23 crc kubenswrapper[4824]: W1124 13:14:23.069058 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:23 crc kubenswrapper[4824]: E1124 13:14:23.069123 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:23 crc kubenswrapper[4824]: W1124 13:14:23.323381 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:23 crc kubenswrapper[4824]: E1124 13:14:23.323467 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:23 crc kubenswrapper[4824]: I1124 13:14:23.940795 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.069496 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9ef690ee36b610c5faddc407465dd0bb33162e56bc5f9ecf04567836d1eb5b20"} Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.069542 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.070499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.070687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.070882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.073426 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8" exitCode=0 Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.073508 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.073511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8"} Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.073532 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.073694 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.074143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.074168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.074176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.074728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.074903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.075003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.205979 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.565285 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.565574 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.567401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.567460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.567481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:24 crc kubenswrapper[4824]: I1124 13:14:24.940557 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:24 crc kubenswrapper[4824]: E1124 13:14:24.993192 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.188:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187af39cd24b6598 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 13:14:18.937574808 +0000 UTC m=+0.577114118,LastTimestamp:2025-11-24 13:14:18.937574808 +0000 UTC m=+0.577114118,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.078996 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c"} Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.079111 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.079125 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.079196 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.080652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.080683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.080686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.080693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.080709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.080720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:25 crc kubenswrapper[4824]: E1124 13:14:25.166580 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.188:6443: connect: connection refused" interval="6.4s" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.229619 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 13:14:25 crc kubenswrapper[4824]: E1124 13:14:25.230859 4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.410480 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.411969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.412014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.412125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.412166 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:14:25 crc kubenswrapper[4824]: E1124 13:14:25.412670 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.188:6443: connect: connection refused" node="crc" Nov 24 13:14:25 crc kubenswrapper[4824]: I1124 13:14:25.940875 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.083725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42"} Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.084338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c"} Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.084356 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f"} Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.084938 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.086286 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9ef690ee36b610c5faddc407465dd0bb33162e56bc5f9ecf04567836d1eb5b20" exitCode=255 Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.086318 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9ef690ee36b610c5faddc407465dd0bb33162e56bc5f9ecf04567836d1eb5b20"} Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.086487 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.087371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.087431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.087444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.088090 4824 scope.go:117] "RemoveContainer" containerID="9ef690ee36b610c5faddc407465dd0bb33162e56bc5f9ecf04567836d1eb5b20" Nov 24 13:14:26 crc kubenswrapper[4824]: W1124 13:14:26.421659 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:26 crc kubenswrapper[4824]: E1124 13:14:26.421771 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.188:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.867193 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:26 crc kubenswrapper[4824]: I1124 13:14:26.941468 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.188:6443: connect: connection refused Nov 24 13:14:27 crc kubenswrapper[4824]: I1124 13:14:27.090675 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 13:14:27 crc kubenswrapper[4824]: I1124 13:14:27.092359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97"} Nov 24 13:14:27 crc kubenswrapper[4824]: I1124 13:14:27.094344 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:27 crc kubenswrapper[4824]: I1124 13:14:27.095821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86"} Nov 24 13:14:27 crc kubenswrapper[4824]: I1124 13:14:27.096108 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:27 crc kubenswrapper[4824]: I1124 13:14:27.097401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:27 crc kubenswrapper[4824]: I1124 13:14:27.097447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:27 crc kubenswrapper[4824]: I1124 13:14:27.097465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:27 crc kubenswrapper[4824]: I1124 13:14:27.939351 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.153711 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.153711 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.154859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.154898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.154912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.155937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.155985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.155997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.236660 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.236873 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.238094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.238144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.238155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:28 crc kubenswrapper[4824]: I1124 13:14:28.426063 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:29 crc kubenswrapper[4824]: E1124 13:14:29.087159 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.156030 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.156101 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.156030 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.157001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.157022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.157030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.157084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.157112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.157125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.910208 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.910442 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.911892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.911946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:29 crc kubenswrapper[4824]: I1124 13:14:29.911957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.020225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.158663 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.158827 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.160620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.160677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.160686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.160889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.160917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.160930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:30 crc kubenswrapper[4824]: I1124 13:14:30.163376 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:31 crc kubenswrapper[4824]: I1124 13:14:31.162150 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:31 crc kubenswrapper[4824]: I1124 13:14:31.164457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:31 crc kubenswrapper[4824]: I1124 13:14:31.164524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:31 crc kubenswrapper[4824]: I1124 13:14:31.164537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:31 crc kubenswrapper[4824]: I1124 13:14:31.813264 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:31 crc kubenswrapper[4824]: I1124 13:14:31.814530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:31 crc kubenswrapper[4824]: I1124 13:14:31.814577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:31 crc kubenswrapper[4824]: I1124 13:14:31.814588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:31 crc kubenswrapper[4824]: I1124 13:14:31.814618 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:14:33 crc kubenswrapper[4824]: I1124 13:14:33.020767 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 13:14:33 crc kubenswrapper[4824]: I1124 13:14:33.020881 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 13:14:33 crc kubenswrapper[4824]: I1124 13:14:33.191842 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 13:14:33 crc kubenswrapper[4824]: I1124 13:14:33.192041 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:33 crc kubenswrapper[4824]: I1124 13:14:33.193024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:33 crc kubenswrapper[4824]: I1124 13:14:33.193058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:33 crc kubenswrapper[4824]: I1124 13:14:33.193069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:33 crc kubenswrapper[4824]: I1124 13:14:33.271875 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 24 13:14:34 crc kubenswrapper[4824]: I1124 13:14:34.888241 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 13:14:34 crc kubenswrapper[4824]: I1124 13:14:34.888306 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 13:14:34 crc kubenswrapper[4824]: I1124 13:14:34.897190 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 24 13:14:34 crc kubenswrapper[4824]: I1124 13:14:34.897258 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.095428 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.095506 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.946865 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.947079 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.947745 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.947872 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.948719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.948908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.949026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:37 crc kubenswrapper[4824]: I1124 13:14:37.952518 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:38 crc kubenswrapper[4824]: I1124 13:14:38.180262 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:38 crc kubenswrapper[4824]: I1124 13:14:38.180754 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 13:14:38 crc kubenswrapper[4824]: I1124 13:14:38.180852 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 13:14:38 crc kubenswrapper[4824]: I1124 13:14:38.181706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:38 crc kubenswrapper[4824]: I1124 13:14:38.181743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:38 crc kubenswrapper[4824]: I1124 13:14:38.181760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:39 crc kubenswrapper[4824]: E1124 13:14:39.087475 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.123395 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.123479 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.873364 4824 trace.go:236] Trace[602446139]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:14:27.227) (total time: 12645ms): Nov 24 13:14:39 crc kubenswrapper[4824]: Trace[602446139]: ---"Objects listed" error: 12645ms (13:14:39.873) Nov 24 13:14:39 crc kubenswrapper[4824]: Trace[602446139]: [12.64591267s] [12.64591267s] END Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.873409 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.873751 4824 trace.go:236] Trace[1279220218]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:14:29.471) (total time: 10401ms): Nov 24 13:14:39 crc kubenswrapper[4824]: Trace[1279220218]: ---"Objects listed" error: 10401ms (13:14:39.873) Nov 24 13:14:39 crc kubenswrapper[4824]: Trace[1279220218]: [10.401836794s] [10.401836794s] END Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.873765 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.874483 4824 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.874564 4824 trace.go:236] Trace[1240918814]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:14:29.262) (total time: 10612ms): Nov 24 13:14:39 crc kubenswrapper[4824]: Trace[1240918814]: ---"Objects listed" error: 10612ms (13:14:39.874) Nov 24 13:14:39 crc kubenswrapper[4824]: Trace[1240918814]: [10.612387889s] [10.612387889s] END Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.874591 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.875146 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 13:14:39 crc kubenswrapper[4824]: E1124 13:14:39.877576 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.910929 4824 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.935214 4824 apiserver.go:52] "Watching apiserver" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.940257 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.940675 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.941134 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.941289 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.941399 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:39 crc kubenswrapper[4824]: E1124 13:14:39.941505 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:39 crc kubenswrapper[4824]: E1124 13:14:39.941575 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.941699 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.942127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.942186 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:14:39 crc kubenswrapper[4824]: E1124 13:14:39.942190 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.943864 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.943950 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.944132 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.944304 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.947293 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.947881 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.948120 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.948534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.953683 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.974663 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.974721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.974753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.974781 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.974853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.974882 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.974920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.974942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.974968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.975555 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.977208 4824 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.978387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.989937 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.990957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:14:39 crc kubenswrapper[4824]: I1124 13:14:39.996903 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.000240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.003767 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.003829 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.003848 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.003960 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:40.503928273 +0000 UTC m=+22.143467583 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.013356 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.015033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.016852 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.016885 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.016903 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.016968 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:40.516947264 +0000 UTC m=+22.156486574 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.023483 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.031798 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.042480 4824 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.053412 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.054940 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076281 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076347 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076379 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076515 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076551 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076574 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076594 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076620 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076644 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076688 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076713 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076736 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076738 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076752 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076758 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076796 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076838 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076865 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076888 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076905 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076922 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076938 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076972 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076991 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077009 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077027 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077045 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077062 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077080 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077134 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077151 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077167 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077199 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077218 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077233 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077250 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077267 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077283 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077326 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077799 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077857 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077880 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077921 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077945 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077966 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077986 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078004 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078030 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078050 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078093 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078114 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078133 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078154 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078171 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078189 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078252 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078270 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078287 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078305 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078322 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078343 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078359 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078378 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078398 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078435 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078453 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078474 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078494 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078510 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078546 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078564 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078581 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078598 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078614 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078631 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078669 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078686 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078705 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078741 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078757 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078776 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078793 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078823 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078841 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078873 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078890 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078906 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078923 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078939 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078990 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079009 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079026 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079043 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079060 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079076 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079093 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079128 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079160 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079198 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079233 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079265 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079282 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079300 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079333 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079349 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079386 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079403 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079438 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079455 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079474 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079496 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079513 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079536 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079555 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079574 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079607 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079623 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079638 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079655 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079674 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079713 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079732 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079751 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079778 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079798 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079844 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079859 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079878 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079927 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079943 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079961 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079979 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080012 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080045 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080062 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080078 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080096 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080113 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080131 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080152 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080169 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080187 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080203 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080222 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080261 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080320 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076990 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.076997 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.077733 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078284 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078358 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078444 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078613 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078747 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.078988 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079130 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079360 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079490 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079524 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079532 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079686 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079689 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.079707 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080162 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080189 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080364 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.080623 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:40.58059404 +0000 UTC m=+22.220133550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080924 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081308 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081459 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081615 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.080373 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081683 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081708 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081732 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081756 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081781 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081824 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081848 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081875 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081900 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081923 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081948 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.081974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082000 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082050 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082073 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082096 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082122 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082145 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082169 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082196 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082220 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082312 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082404 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082680 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082698 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082712 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082726 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082740 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082753 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082765 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082779 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082791 4824 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082863 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082880 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082892 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082904 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082917 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082929 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082941 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082956 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082969 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.082984 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083000 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083013 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083026 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083039 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083051 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083064 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083077 4824 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083090 4824 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083103 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083115 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.083130 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.089169 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.089334 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.089643 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.089817 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.090044 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.090926 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.091056 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.091214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.092006 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.092276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.092294 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.092533 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.093374 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.094045 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.094370 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.094852 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.095320 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.095580 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.095819 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.096014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.096367 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.096641 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.096861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.097492 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.097607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.098390 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.098391 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.098604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.098665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.098674 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.098800 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.098969 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.099035 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.099028 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.099279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.099305 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.099534 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.099539 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.099553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.099972 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.099974 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.100111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.100667 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.101007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.101475 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.101673 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.101692 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.102108 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.102972 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.102996 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.102978 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.103121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.103388 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.103413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.103511 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.104074 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.104096 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.104201 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.104355 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.104617 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.104661 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.104794 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105023 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.105110 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105146 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.105204 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:40.605182813 +0000 UTC m=+22.244722123 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105383 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105530 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105602 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105785 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105960 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106044 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106129 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106925 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.105989 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106207 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106298 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106358 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106508 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106621 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106657 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.106666 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.107173 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.107301 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.107359 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:40.607349591 +0000 UTC m=+22.246888901 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.107382 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.107424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.108977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.109356 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.109673 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.110051 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.110089 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.113106 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.113139 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.115401 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.115931 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.115938 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.116128 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.116252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.116769 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.116259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.116333 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.116539 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.116882 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.117025 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.116762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.117155 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.117278 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.117296 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.117760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.117975 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.118170 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.118473 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.118794 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.118860 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.118898 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.119021 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.119273 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.119386 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.119895 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.119990 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.120471 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.120507 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.120155 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.120396 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.121044 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.121293 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.121434 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.122371 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.123149 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.123248 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.123836 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.124345 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.124545 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.124562 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.124626 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.127142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.127192 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.127502 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.127929 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.128866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.129570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.129698 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.129772 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.129882 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.130061 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.130078 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.130213 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.130321 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.130474 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.130530 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.130566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.138781 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.139254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.139609 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.140001 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.140058 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.140500 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.144133 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.143945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.144340 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.144516 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.144910 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.145211 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.145421 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.148597 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.148838 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.155278 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.161413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.164012 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.165098 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.177115 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.177026 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.180789 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184296 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184419 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184442 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184457 4824 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184496 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184513 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184525 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184539 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184583 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184599 4824 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184611 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184624 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184652 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184663 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184673 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184683 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184694 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184705 4824 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184735 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184746 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184755 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184765 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184778 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184787 4824 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184833 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184853 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184867 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184899 4824 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184909 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184919 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184931 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184942 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.184999 4824 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185009 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185068 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185078 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185087 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185096 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185108 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185117 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185145 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185170 4824 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185180 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185190 4824 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185223 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185233 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185243 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185252 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185261 4824 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185270 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185300 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185309 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185319 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185329 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185340 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185349 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185358 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185366 4824 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185376 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185384 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185394 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185402 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185412 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185420 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185429 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185440 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185449 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185458 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185467 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185476 4824 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185486 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185494 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185505 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185517 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185531 4824 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185543 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185555 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185566 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185575 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185589 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185599 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185610 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185623 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185634 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185645 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185655 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185666 4824 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185677 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185688 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185698 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185709 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185729 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185741 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185752 4824 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185763 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185773 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185784 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185901 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185935 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185948 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185959 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185970 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185982 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.185992 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186003 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186034 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186143 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186154 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186164 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186175 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186185 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186213 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186223 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186232 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186240 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186249 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186258 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186270 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186323 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186336 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186345 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186354 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186389 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186401 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186411 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186420 4824 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186428 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186437 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186467 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186477 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186489 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186497 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186506 4824 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186515 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186525 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186537 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186588 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186598 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186644 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186655 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186664 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186673 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186682 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186691 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186700 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186711 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186720 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186729 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186737 4824 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186747 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186756 4824 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186765 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186775 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186784 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186793 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186814 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186824 4824 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186833 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186843 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186852 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186862 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186871 4824 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186884 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.186896 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.188593 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.196452 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.203386 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.215541 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.224831 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.234919 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.244689 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.264170 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.272590 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.281036 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:14:40 crc kubenswrapper[4824]: W1124 13:14:40.281634 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-cc6296080aa45a60fa7c9a8f1a88499b2a1a8d9fef4c704d145b8ead449088ac WatchSource:0}: Error finding container cc6296080aa45a60fa7c9a8f1a88499b2a1a8d9fef4c704d145b8ead449088ac: Status 404 returned error can't find the container with id cc6296080aa45a60fa7c9a8f1a88499b2a1a8d9fef4c704d145b8ead449088ac Nov 24 13:14:40 crc kubenswrapper[4824]: W1124 13:14:40.293223 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-543dfbe362a2bf427e9c74f244bbcc0da7a84367d5080ba84198f60bc34dbcc9 WatchSource:0}: Error finding container 543dfbe362a2bf427e9c74f244bbcc0da7a84367d5080ba84198f60bc34dbcc9: Status 404 returned error can't find the container with id 543dfbe362a2bf427e9c74f244bbcc0da7a84367d5080ba84198f60bc34dbcc9 Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.591229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.591439 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:41.591407953 +0000 UTC m=+23.230947263 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.591664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.591699 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.591875 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.591894 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.591907 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.591966 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:41.591945957 +0000 UTC m=+23.231485267 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.592034 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.592046 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.592054 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.592079 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:41.59207137 +0000 UTC m=+23.231610680 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.693002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:40 crc kubenswrapper[4824]: I1124 13:14:40.693048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.693163 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.693218 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:41.693202557 +0000 UTC m=+23.332741867 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.693245 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:40 crc kubenswrapper[4824]: E1124 13:14:40.693484 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:41.693437904 +0000 UTC m=+23.332977214 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.013765 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.014294 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.015605 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.016231 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.017203 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.017670 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.018263 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.019201 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.019777 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.020723 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.021370 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.022632 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.023170 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.023726 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.024740 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.025343 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.026404 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.026873 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.027414 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.028436 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.028929 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.029843 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.030267 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.031233 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.031652 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.032273 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.033450 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.033905 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.034794 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.035247 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.036082 4824 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.036181 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.037787 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.038823 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.039223 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.040680 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.041349 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.042343 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.043034 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.043773 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.044299 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.044949 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.045608 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.046258 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.046690 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.047223 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.047735 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.048421 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.048933 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.049406 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.049915 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.050475 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.051186 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.055210 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.191175 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"543dfbe362a2bf427e9c74f244bbcc0da7a84367d5080ba84198f60bc34dbcc9"} Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.199572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d"} Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.199678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a"} Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.199691 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d8e18c6398edd0c5b0005bc34aea90534d07baf64684eec9675316c5a1dafbc4"} Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.201536 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76"} Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.201589 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"cc6296080aa45a60fa7c9a8f1a88499b2a1a8d9fef4c704d145b8ead449088ac"} Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.203328 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.203949 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.206540 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97" exitCode=255 Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.206591 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97"} Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.206713 4824 scope.go:117] "RemoveContainer" containerID="9ef690ee36b610c5faddc407465dd0bb33162e56bc5f9ecf04567836d1eb5b20" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.220039 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.221932 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.222098 4824 scope.go:117] "RemoveContainer" containerID="3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97" Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.222346 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.240023 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.257257 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.270893 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.288127 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.303919 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.319449 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.336325 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ef690ee36b610c5faddc407465dd0bb33162e56bc5f9ecf04567836d1eb5b20\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:25Z\\\",\\\"message\\\":\\\"W1124 13:14:24.455240 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 13:14:24.455602 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763990064 cert, and key in /tmp/serving-cert-4133978030/serving-signer.crt, /tmp/serving-cert-4133978030/serving-signer.key\\\\nI1124 13:14:25.071163 1 observer_polling.go:159] Starting file observer\\\\nW1124 13:14:25.083198 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 13:14:25.083356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:25.084107 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4133978030/tls.crt::/tmp/serving-cert-4133978030/tls.key\\\\\\\"\\\\nF1124 13:14:25.398170 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.350069 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.362273 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.373392 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.387023 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.398873 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.410855 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.426787 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:41Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.601323 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.601476 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.601584 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:43.601540528 +0000 UTC m=+25.241079838 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.601635 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.601660 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.601679 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.601724 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.601748 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:43.601727913 +0000 UTC m=+25.241267233 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.601910 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.601929 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.601941 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.601991 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:43.60198227 +0000 UTC m=+25.241521570 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.702976 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:41 crc kubenswrapper[4824]: I1124 13:14:41.703037 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.703126 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.703145 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.703207 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:43.703186889 +0000 UTC m=+25.342726199 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:41 crc kubenswrapper[4824]: E1124 13:14:41.703226 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:43.70321987 +0000 UTC m=+25.342759180 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.009356 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.009400 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.009367 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:42 crc kubenswrapper[4824]: E1124 13:14:42.009553 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:42 crc kubenswrapper[4824]: E1124 13:14:42.009669 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:42 crc kubenswrapper[4824]: E1124 13:14:42.009735 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.211858 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.215079 4824 scope.go:117] "RemoveContainer" containerID="3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97" Nov 24 13:14:42 crc kubenswrapper[4824]: E1124 13:14:42.215222 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.235320 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.276827 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.297484 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.326115 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.349180 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.378459 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.394207 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:42 crc kubenswrapper[4824]: I1124 13:14:42.409106 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.219193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073"} Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.220577 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.235046 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.236699 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.241589 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.257198 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.274864 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.294780 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.310750 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.324578 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.339557 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.355093 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.372222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.387617 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.400141 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.416336 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.433769 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.447708 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.474834 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.491446 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.505457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:43Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.620615 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.620695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.620717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.620889 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:47.620850983 +0000 UTC m=+29.260390293 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.620893 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.621018 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.621038 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.621103 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:47.621091959 +0000 UTC m=+29.260631269 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.621108 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.621177 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.621194 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.621279 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:47.621249984 +0000 UTC m=+29.260789334 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.721564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:43 crc kubenswrapper[4824]: I1124 13:14:43.721643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.721721 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.721784 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:47.721766794 +0000 UTC m=+29.361306104 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.721779 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:43 crc kubenswrapper[4824]: E1124 13:14:43.721932 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:47.721906258 +0000 UTC m=+29.361445598 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:44 crc kubenswrapper[4824]: I1124 13:14:44.009300 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:44 crc kubenswrapper[4824]: I1124 13:14:44.009358 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:44 crc kubenswrapper[4824]: I1124 13:14:44.009308 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:44 crc kubenswrapper[4824]: E1124 13:14:44.009530 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:44 crc kubenswrapper[4824]: E1124 13:14:44.009454 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:44 crc kubenswrapper[4824]: E1124 13:14:44.009680 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:44 crc kubenswrapper[4824]: E1124 13:14:44.244224 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Nov 24 13:14:45 crc kubenswrapper[4824]: I1124 13:14:45.534862 4824 csr.go:261] certificate signing request csr-lhhnt is approved, waiting to be issued Nov 24 13:14:45 crc kubenswrapper[4824]: I1124 13:14:45.552941 4824 csr.go:257] certificate signing request csr-lhhnt is issued Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.010030 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.010106 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:46 crc kubenswrapper[4824]: E1124 13:14:46.010200 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:46 crc kubenswrapper[4824]: E1124 13:14:46.010395 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.010538 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:46 crc kubenswrapper[4824]: E1124 13:14:46.010764 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.044005 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-xtlj7"] Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.044369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xtlj7" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.046875 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.047833 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.048211 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.062461 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.077021 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.092133 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.110020 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.133567 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.142962 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4-hosts-file\") pod \"node-resolver-xtlj7\" (UID: \"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\") " pod="openshift-dns/node-resolver-xtlj7" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.143286 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mpbm\" (UniqueName: \"kubernetes.io/projected/d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4-kube-api-access-6mpbm\") pod \"node-resolver-xtlj7\" (UID: \"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\") " pod="openshift-dns/node-resolver-xtlj7" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.164242 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.193451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.233429 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.244611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4-hosts-file\") pod \"node-resolver-xtlj7\" (UID: \"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\") " pod="openshift-dns/node-resolver-xtlj7" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.244653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mpbm\" (UniqueName: \"kubernetes.io/projected/d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4-kube-api-access-6mpbm\") pod \"node-resolver-xtlj7\" (UID: \"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\") " pod="openshift-dns/node-resolver-xtlj7" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.245005 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4-hosts-file\") pod \"node-resolver-xtlj7\" (UID: \"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\") " pod="openshift-dns/node-resolver-xtlj7" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.266363 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.267017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mpbm\" (UniqueName: \"kubernetes.io/projected/d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4-kube-api-access-6mpbm\") pod \"node-resolver-xtlj7\" (UID: \"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\") " pod="openshift-dns/node-resolver-xtlj7" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.293897 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.359730 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xtlj7" Nov 24 13:14:46 crc kubenswrapper[4824]: W1124 13:14:46.375854 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5f9693d_cfd5_4c65_af6a_45d54eb2f5c4.slice/crio-37a495936b15ac992871027d3e0100bd29c5f323fe18b8913c2638c67cdd3c03 WatchSource:0}: Error finding container 37a495936b15ac992871027d3e0100bd29c5f323fe18b8913c2638c67cdd3c03: Status 404 returned error can't find the container with id 37a495936b15ac992871027d3e0100bd29c5f323fe18b8913c2638c67cdd3c03 Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.483587 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-9wxxt"] Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.484224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.485180 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-ssstr"] Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.485552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.485677 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-74f4n"] Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.486169 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.516055 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.516065 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.516302 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.516367 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.522312 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.522525 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.522655 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.523252 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.523454 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.523723 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.523772 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.526299 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547124 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzlnp\" (UniqueName: \"kubernetes.io/projected/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-kube-api-access-lzlnp\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547246 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-var-lib-cni-multus\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547279 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-multus-cni-dir\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-var-lib-cni-bin\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-var-lib-kubelet\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-run-netns\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547434 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-cnibin\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547518 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-run-k8s-cni-cncf-io\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547562 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfdfb\" (UniqueName: \"kubernetes.io/projected/b90f50d7-5574-472e-9532-91b491a0cc56-kube-api-access-bfdfb\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547615 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-os-release\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2599a7bb-fad5-4c40-901b-189dc6687928-cni-binary-copy\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwlhd\" (UniqueName: \"kubernetes.io/projected/2599a7bb-fad5-4c40-901b-189dc6687928-kube-api-access-lwlhd\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547668 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-system-cni-dir\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547682 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-multus-socket-dir-parent\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547721 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-cnibin\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-cni-binary-copy\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547755 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-os-release\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b90f50d7-5574-472e-9532-91b491a0cc56-rootfs\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b90f50d7-5574-472e-9532-91b491a0cc56-mcd-auth-proxy-config\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547863 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-etc-kubernetes\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547879 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-multus-conf-dir\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547895 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2599a7bb-fad5-4c40-901b-189dc6687928-multus-daemon-config\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547912 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b90f50d7-5574-472e-9532-91b491a0cc56-proxy-tls\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547929 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-hostroot\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547945 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-run-multus-certs\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.547959 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-system-cni-dir\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.551592 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.554359 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-24 13:09:45 +0000 UTC, rotation deadline is 2026-08-09 22:11:00.712146141 +0000 UTC Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.554416 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6200h56m14.157731851s for next certificate rotation Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.590396 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.625233 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.641564 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649618 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649655 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-os-release\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649674 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b90f50d7-5574-472e-9532-91b491a0cc56-mcd-auth-proxy-config\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b90f50d7-5574-472e-9532-91b491a0cc56-rootfs\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-etc-kubernetes\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-multus-conf-dir\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649782 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2599a7bb-fad5-4c40-901b-189dc6687928-multus-daemon-config\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649797 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b90f50d7-5574-472e-9532-91b491a0cc56-proxy-tls\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-system-cni-dir\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-hostroot\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649860 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-run-multus-certs\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzlnp\" (UniqueName: \"kubernetes.io/projected/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-kube-api-access-lzlnp\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649903 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-var-lib-cni-multus\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649940 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-multus-cni-dir\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649956 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-var-lib-cni-bin\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649972 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-var-lib-kubelet\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.649996 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-run-netns\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650014 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-cnibin\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-run-k8s-cni-cncf-io\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfdfb\" (UniqueName: \"kubernetes.io/projected/b90f50d7-5574-472e-9532-91b491a0cc56-kube-api-access-bfdfb\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650065 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwlhd\" (UniqueName: \"kubernetes.io/projected/2599a7bb-fad5-4c40-901b-189dc6687928-kube-api-access-lwlhd\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650081 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-os-release\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2599a7bb-fad5-4c40-901b-189dc6687928-cni-binary-copy\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650116 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-cnibin\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-cni-binary-copy\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-system-cni-dir\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650163 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-multus-socket-dir-parent\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650351 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-multus-socket-dir-parent\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.650946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651092 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-os-release\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651529 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-multus-cni-dir\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651568 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b90f50d7-5574-472e-9532-91b491a0cc56-mcd-auth-proxy-config\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b90f50d7-5574-472e-9532-91b491a0cc56-rootfs\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651613 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-var-lib-cni-bin\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651625 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-etc-kubernetes\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651638 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-var-lib-kubelet\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651661 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-multus-conf-dir\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651664 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-run-netns\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651691 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-cnibin\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.651714 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-run-k8s-cni-cncf-io\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652031 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-system-cni-dir\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652160 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-cnibin\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652211 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-os-release\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-run-multus-certs\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-system-cni-dir\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652311 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-hostroot\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652357 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2599a7bb-fad5-4c40-901b-189dc6687928-multus-daemon-config\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652414 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2599a7bb-fad5-4c40-901b-189dc6687928-host-var-lib-cni-multus\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2599a7bb-fad5-4c40-901b-189dc6687928-cni-binary-copy\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.652737 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-cni-binary-copy\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.653499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.656927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b90f50d7-5574-472e-9532-91b491a0cc56-proxy-tls\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.664152 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.667902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfdfb\" (UniqueName: \"kubernetes.io/projected/b90f50d7-5574-472e-9532-91b491a0cc56-kube-api-access-bfdfb\") pod \"machine-config-daemon-74f4n\" (UID: \"b90f50d7-5574-472e-9532-91b491a0cc56\") " pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.668476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzlnp\" (UniqueName: \"kubernetes.io/projected/fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0-kube-api-access-lzlnp\") pod \"multus-additional-cni-plugins-9wxxt\" (UID: \"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\") " pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.670427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwlhd\" (UniqueName: \"kubernetes.io/projected/2599a7bb-fad5-4c40-901b-189dc6687928-kube-api-access-lwlhd\") pod \"multus-ssstr\" (UID: \"2599a7bb-fad5-4c40-901b-189dc6687928\") " pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.682114 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.694321 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.712887 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.727868 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.742853 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.758923 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.773702 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.787961 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.799330 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.803621 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.806764 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ssstr" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.813742 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:14:46 crc kubenswrapper[4824]: W1124 13:14:46.815147 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc1fdc8e_08cd_4deb_b00c_8b84f656a6d0.slice/crio-4934f77729d131f9a620fff1969e6a90c42a90258eb2ce495f2b9ed6765a77b7 WatchSource:0}: Error finding container 4934f77729d131f9a620fff1969e6a90c42a90258eb2ce495f2b9ed6765a77b7: Status 404 returned error can't find the container with id 4934f77729d131f9a620fff1969e6a90c42a90258eb2ce495f2b9ed6765a77b7 Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.829201 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: W1124 13:14:46.833602 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2599a7bb_fad5_4c40_901b_189dc6687928.slice/crio-671a59698e7dc31d2d1899318b4a13a2252a62c9390bec9c57ea874ab44e94f5 WatchSource:0}: Error finding container 671a59698e7dc31d2d1899318b4a13a2252a62c9390bec9c57ea874ab44e94f5: Status 404 returned error can't find the container with id 671a59698e7dc31d2d1899318b4a13a2252a62c9390bec9c57ea874ab44e94f5 Nov 24 13:14:46 crc kubenswrapper[4824]: W1124 13:14:46.840965 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb90f50d7_5574_472e_9532_91b491a0cc56.slice/crio-1a823164fcfcdf095cc5347102f83ac8b9ba85dcea5980feb639cf11f12eda1b WatchSource:0}: Error finding container 1a823164fcfcdf095cc5347102f83ac8b9ba85dcea5980feb639cf11f12eda1b: Status 404 returned error can't find the container with id 1a823164fcfcdf095cc5347102f83ac8b9ba85dcea5980feb639cf11f12eda1b Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.846173 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.878720 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.881564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.881601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.881610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.881725 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.886092 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.912619 4824 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.912929 4824 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.917495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.917609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.917689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.917762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.917992 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:46Z","lastTransitionTime":"2025-11-24T13:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.925929 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nk5qz"] Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.926216 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.927442 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.933500 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.933551 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.933673 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.933827 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.934257 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.934497 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.934635 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 13:14:46 crc kubenswrapper[4824]: E1124 13:14:46.955484 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.962896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.962934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.962945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.962961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.962971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:46Z","lastTransitionTime":"2025-11-24T13:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.968738 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: E1124 13:14:46.980365 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.990257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.990293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.990304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.990320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.990329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:46Z","lastTransitionTime":"2025-11-24T13:14:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:46 crc kubenswrapper[4824]: I1124 13:14:46.998397 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.011547 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.015407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.015446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.015457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.015473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.015484 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.016288 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.033216 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.035910 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.038363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.038385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.038396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.038413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.038424 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.051205 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.055358 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.055560 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-kubelet\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058325 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-var-lib-openvswitch\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058355 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-systemd\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058381 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-ovn-kubernetes\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058399 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-systemd-units\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058417 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-log-socket\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-env-overrides\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058459 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vgzk\" (UniqueName: \"kubernetes.io/projected/0f43932f-b763-4fea-9ed9-737078e6548d-kube-api-access-8vgzk\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-netns\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-etc-openvswitch\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-openvswitch\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058595 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-ovn\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-script-lib\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058638 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-config\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-bin\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f43932f-b763-4fea-9ed9-737078e6548d-ovn-node-metrics-cert\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058686 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-slash\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058719 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-node-log\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.058747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-netd\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.061020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.061059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.061075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.061094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.061106 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.069703 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.085534 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.099292 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.112344 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.127336 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.145434 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.159773 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-config\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.159857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-bin\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.159880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f43932f-b763-4fea-9ed9-737078e6548d-ovn-node-metrics-cert\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.159909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-slash\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.159930 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-node-log\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.159946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-netd\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.159964 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-kubelet\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.159984 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-var-lib-openvswitch\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-systemd\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-ovn-kubernetes\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160099 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-systemd-units\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-log-socket\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160145 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-env-overrides\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160344 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-bin\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160508 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vgzk\" (UniqueName: \"kubernetes.io/projected/0f43932f-b763-4fea-9ed9-737078e6548d-kube-api-access-8vgzk\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-node-log\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160671 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-var-lib-openvswitch\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160736 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-slash\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160785 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-netd\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160843 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-config\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-ovn-kubernetes\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-systemd\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160923 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-kubelet\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-systemd-units\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.160968 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-log-socket\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.161010 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-env-overrides\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.161153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.161194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-netns\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.161219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-etc-openvswitch\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.161238 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-openvswitch\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.161309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-ovn\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.161334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-script-lib\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.161435 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-openvswitch\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.161332 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.162938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-script-lib\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.162016 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.162998 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-ovn\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.162002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-netns\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.163064 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-etc-openvswitch\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.167989 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f43932f-b763-4fea-9ed9-737078e6548d-ovn-node-metrics-cert\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.169190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.169225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.169234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.169249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.169264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.191077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vgzk\" (UniqueName: \"kubernetes.io/projected/0f43932f-b763-4fea-9ed9-737078e6548d-kube-api-access-8vgzk\") pod \"ovnkube-node-nk5qz\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.199787 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.215009 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.229931 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.231966 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.232025 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.232036 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"1a823164fcfcdf095cc5347102f83ac8b9ba85dcea5980feb639cf11f12eda1b"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.233336 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xtlj7" event={"ID":"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4","Type":"ContainerStarted","Data":"a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.233382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xtlj7" event={"ID":"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4","Type":"ContainerStarted","Data":"37a495936b15ac992871027d3e0100bd29c5f323fe18b8913c2638c67cdd3c03"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.235169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ssstr" event={"ID":"2599a7bb-fad5-4c40-901b-189dc6687928","Type":"ContainerStarted","Data":"42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.235237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ssstr" event={"ID":"2599a7bb-fad5-4c40-901b-189dc6687928","Type":"ContainerStarted","Data":"671a59698e7dc31d2d1899318b4a13a2252a62c9390bec9c57ea874ab44e94f5"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.236316 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerStarted","Data":"e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.236340 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerStarted","Data":"4934f77729d131f9a620fff1969e6a90c42a90258eb2ce495f2b9ed6765a77b7"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.244927 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.260394 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.267051 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.271845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.271978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.272141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.272231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.272295 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.282364 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.296872 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.310717 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.332826 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.347305 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.362111 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.374497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.374538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.374552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.374571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.374582 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.377370 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.393022 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.416540 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.434760 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.449282 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.462287 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.477730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.477842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.477868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.477891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.477932 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.479782 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.492476 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.513083 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.527667 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.543009 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.580896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.580950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.580963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.580985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.581005 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.667252 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.667474 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.667446769 +0000 UTC m=+37.306986089 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.667628 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.667901 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.667930 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.667945 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.667994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.668038 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.668048 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.668058 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.668087 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.668076116 +0000 UTC m=+37.307615426 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.668174 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.668158358 +0000 UTC m=+37.307697668 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.684438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.684481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.684496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.684518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.684530 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.768770 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.768894 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.768865753 +0000 UTC m=+37.408405073 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.768980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.769070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.769332 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:47 crc kubenswrapper[4824]: E1124 13:14:47.769375 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:14:55.769359927 +0000 UTC m=+37.408899237 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.787798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.787868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.787886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.787904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.787916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.890909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.890948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.890963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.891029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.891046 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.993550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.993599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.993610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.993627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:47 crc kubenswrapper[4824]: I1124 13:14:47.993640 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:47Z","lastTransitionTime":"2025-11-24T13:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.009882 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.010028 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.010155 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:48 crc kubenswrapper[4824]: E1124 13:14:48.010233 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:48 crc kubenswrapper[4824]: E1124 13:14:48.010294 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:48 crc kubenswrapper[4824]: E1124 13:14:48.010367 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.098659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.098714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.098728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.098790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.098828 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:48Z","lastTransitionTime":"2025-11-24T13:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.201724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.201769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.201782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.201829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.201880 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:48Z","lastTransitionTime":"2025-11-24T13:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.305157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.305216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.305229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.305253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.305268 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:48Z","lastTransitionTime":"2025-11-24T13:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.408201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.408232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.408241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.408256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.408265 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:48Z","lastTransitionTime":"2025-11-24T13:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.510197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.510227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.510238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.510251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.510260 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:48Z","lastTransitionTime":"2025-11-24T13:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.612622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.612706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.612721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.612740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.613179 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:48Z","lastTransitionTime":"2025-11-24T13:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.715161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.715205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.715217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.715235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.715246 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:48Z","lastTransitionTime":"2025-11-24T13:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.817225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.817259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.817268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.817284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.817295 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:48Z","lastTransitionTime":"2025-11-24T13:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.821421 4824 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.919597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.919647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.919659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.919675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:48 crc kubenswrapper[4824]: I1124 13:14:48.919685 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:48Z","lastTransitionTime":"2025-11-24T13:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.022130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.022183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.022195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.022227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.022240 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.023374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.033056 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.044433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.054593 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.067140 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.079294 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.107010 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.120136 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.123087 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.124130 4824 scope.go:117] "RemoveContainer" containerID="3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97" Nov 24 13:14:49 crc kubenswrapper[4824]: E1124 13:14:49.124291 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.125003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.125038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.125095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.125113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.125124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.136058 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.151345 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.171696 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.191637 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.206073 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.218828 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.227064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.227108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.227120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.227139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.227207 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.243218 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7" exitCode=0 Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.243304 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.243352 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"e9803f40667b9c49bc7cd16cd9f867edd39641616cbe80297f72d1cfe7e3a25c"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.244452 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0" containerID="e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c" exitCode=0 Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.244480 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerDied","Data":"e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.258367 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.281170 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.298457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.314469 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.326722 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.330042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.330077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.330088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.330106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.330117 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.342327 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.356485 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.382484 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.405381 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.418453 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.432628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.432686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.432712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.432741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.432755 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.435011 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.448969 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.464293 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.487735 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.511587 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.526906 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.535385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.535418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.535428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.535443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.535451 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.545436 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.558706 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.575167 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.590210 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.604583 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.620956 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.638175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.638216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.638226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.638241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.638270 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.657977 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.673833 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.689914 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.708794 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.727716 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.741626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.742021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.742034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.742096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.742119 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.749912 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.766978 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-lfmjc"] Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.767517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.774505 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.774792 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.774961 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.775098 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.797947 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.810795 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.823705 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.838372 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.845180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.845213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.845224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.845239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.845250 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.852207 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.864595 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.877841 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.889604 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.896501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t45j\" (UniqueName: \"kubernetes.io/projected/e986e9a8-5aca-46af-a68c-b871a678fb37-kube-api-access-7t45j\") pod \"node-ca-lfmjc\" (UID: \"e986e9a8-5aca-46af-a68c-b871a678fb37\") " pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.896560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e986e9a8-5aca-46af-a68c-b871a678fb37-host\") pod \"node-ca-lfmjc\" (UID: \"e986e9a8-5aca-46af-a68c-b871a678fb37\") " pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.896582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e986e9a8-5aca-46af-a68c-b871a678fb37-serviceca\") pod \"node-ca-lfmjc\" (UID: \"e986e9a8-5aca-46af-a68c-b871a678fb37\") " pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.900337 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.920104 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.941796 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.947763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.947819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.947833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.947852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.947864 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:49Z","lastTransitionTime":"2025-11-24T13:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.960992 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.974599 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.990157 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.997489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e986e9a8-5aca-46af-a68c-b871a678fb37-host\") pod \"node-ca-lfmjc\" (UID: \"e986e9a8-5aca-46af-a68c-b871a678fb37\") " pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.997589 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e986e9a8-5aca-46af-a68c-b871a678fb37-serviceca\") pod \"node-ca-lfmjc\" (UID: \"e986e9a8-5aca-46af-a68c-b871a678fb37\") " pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.997690 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t45j\" (UniqueName: \"kubernetes.io/projected/e986e9a8-5aca-46af-a68c-b871a678fb37-kube-api-access-7t45j\") pod \"node-ca-lfmjc\" (UID: \"e986e9a8-5aca-46af-a68c-b871a678fb37\") " pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.998944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e986e9a8-5aca-46af-a68c-b871a678fb37-serviceca\") pod \"node-ca-lfmjc\" (UID: \"e986e9a8-5aca-46af-a68c-b871a678fb37\") " pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:49 crc kubenswrapper[4824]: I1124 13:14:49.999047 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e986e9a8-5aca-46af-a68c-b871a678fb37-host\") pod \"node-ca-lfmjc\" (UID: \"e986e9a8-5aca-46af-a68c-b871a678fb37\") " pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.008130 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.010157 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.010215 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.010303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:50 crc kubenswrapper[4824]: E1124 13:14:50.010446 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:50 crc kubenswrapper[4824]: E1124 13:14:50.010580 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:50 crc kubenswrapper[4824]: E1124 13:14:50.010716 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.021448 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t45j\" (UniqueName: \"kubernetes.io/projected/e986e9a8-5aca-46af-a68c-b871a678fb37-kube-api-access-7t45j\") pod \"node-ca-lfmjc\" (UID: \"e986e9a8-5aca-46af-a68c-b871a678fb37\") " pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.051133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.051175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.051187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.051202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.051213 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.154097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.154141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.154151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.154168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.154179 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.202122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lfmjc" Nov 24 13:14:50 crc kubenswrapper[4824]: W1124 13:14:50.216095 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode986e9a8_5aca_46af_a68c_b871a678fb37.slice/crio-6ae75c5e20b3649cad7ca195f071a20d710a44ebd5ab8347be9bec2b583ce2b7 WatchSource:0}: Error finding container 6ae75c5e20b3649cad7ca195f071a20d710a44ebd5ab8347be9bec2b583ce2b7: Status 404 returned error can't find the container with id 6ae75c5e20b3649cad7ca195f071a20d710a44ebd5ab8347be9bec2b583ce2b7 Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.251287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lfmjc" event={"ID":"e986e9a8-5aca-46af-a68c-b871a678fb37","Type":"ContainerStarted","Data":"6ae75c5e20b3649cad7ca195f071a20d710a44ebd5ab8347be9bec2b583ce2b7"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.254266 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.254297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.254307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.256114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.256182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.256194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.256214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.256248 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.258258 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0" containerID="84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3" exitCode=0 Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.258339 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerDied","Data":"84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.271990 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.285036 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.302768 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.331681 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.353522 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.360960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.360980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.360988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.361006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.361015 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.368773 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.382261 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.392788 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.405349 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.421021 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.434649 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.447457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.459402 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.463158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.463184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.463192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.463205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.463214 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.472918 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.492730 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.566418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.566718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.566731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.566746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.566756 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.668961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.668995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.669004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.669020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.669031 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.771896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.771945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.771958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.771981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.772006 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.874409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.874460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.874472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.874492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.874504 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.977623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.977673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.977686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.977707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:50 crc kubenswrapper[4824]: I1124 13:14:50.977723 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:50Z","lastTransitionTime":"2025-11-24T13:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.080746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.081064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.081165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.081247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.081304 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:51Z","lastTransitionTime":"2025-11-24T13:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.188594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.188654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.188670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.188694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.188715 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:51Z","lastTransitionTime":"2025-11-24T13:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.265024 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0" containerID="b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002" exitCode=0 Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.265058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerDied","Data":"b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.268396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lfmjc" event={"ID":"e986e9a8-5aca-46af-a68c-b871a678fb37","Type":"ContainerStarted","Data":"17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.272867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.272926 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.272940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.282979 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.291712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.291737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.291748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.291763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.291773 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:51Z","lastTransitionTime":"2025-11-24T13:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.298687 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.308731 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.318520 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.334884 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.350606 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.367728 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.390873 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.394416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.394454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.394466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.394509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.394520 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:51Z","lastTransitionTime":"2025-11-24T13:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.406457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.421648 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.439052 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.454558 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.477253 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.492271 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.497824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.497847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.497856 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.497870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.497879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:51Z","lastTransitionTime":"2025-11-24T13:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.506346 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.520893 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.535130 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.546612 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.565061 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.577074 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.590011 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.600619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.600675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.600686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.600703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.600717 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:51Z","lastTransitionTime":"2025-11-24T13:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.606911 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.624960 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.646217 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.657997 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.670098 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.687766 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.700739 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.703636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.703712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.703726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.703747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.703759 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:51Z","lastTransitionTime":"2025-11-24T13:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.711522 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.723502 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.806292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.806340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.806351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.806369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.806383 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:51Z","lastTransitionTime":"2025-11-24T13:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.910022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.910086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.910102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.910126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:51 crc kubenswrapper[4824]: I1124 13:14:51.910140 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:51Z","lastTransitionTime":"2025-11-24T13:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.009863 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:52 crc kubenswrapper[4824]: E1124 13:14:52.010157 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.010746 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:52 crc kubenswrapper[4824]: E1124 13:14:52.010937 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.011084 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:52 crc kubenswrapper[4824]: E1124 13:14:52.011262 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.012606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.012638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.012651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.012671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.012684 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.115175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.115219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.115228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.115243 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.115252 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.217899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.217939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.217949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.217967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.217980 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.278233 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0" containerID="32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1" exitCode=0 Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.278327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerDied","Data":"32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.293622 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.309993 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.320514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.320653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.320824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.320899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.320988 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.322885 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.334780 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.346186 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.360629 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.374153 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.385279 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.394957 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.406658 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.420526 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.423542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.423580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.423595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.423613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.423625 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.442366 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.462273 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.475507 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.489400 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.526473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.526511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.526520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.526535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.526545 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.628681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.628715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.628727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.628742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.628751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.731573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.732160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.732240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.732354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.732441 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.835624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.835678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.835696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.835716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.835728 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.938928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.938999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.939016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.939045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:52 crc kubenswrapper[4824]: I1124 13:14:52.939103 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:52Z","lastTransitionTime":"2025-11-24T13:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.041706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.041736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.041744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.041760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.041769 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.144373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.144413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.144423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.144440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.144451 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.247331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.247669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.247679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.247694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.247704 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.283149 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerStarted","Data":"d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.301222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.314852 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.329568 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.350995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.351035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.351045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.351062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.351073 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.357821 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.378579 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.392526 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.405875 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.423732 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.441599 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.453127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.453168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.453176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.453191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.453202 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.456337 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.467439 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.477312 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.491429 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.502491 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.511849 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.555542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.555596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.555609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.555630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.555644 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.658327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.658386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.658400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.658442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.658456 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.761120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.761163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.761174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.761194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.761206 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.864835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.864866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.864876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.864894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.864907 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.967852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.967901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.967921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.967944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:53 crc kubenswrapper[4824]: I1124 13:14:53.967958 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:53Z","lastTransitionTime":"2025-11-24T13:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.009943 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.009941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.010109 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:54 crc kubenswrapper[4824]: E1124 13:14:54.010247 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:54 crc kubenswrapper[4824]: E1124 13:14:54.010423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:54 crc kubenswrapper[4824]: E1124 13:14:54.010522 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.071112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.071168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.071188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.071208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.071220 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.174453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.174494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.174504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.174521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.174534 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.277740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.277774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.277784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.277801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.277825 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.289380 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0" containerID="d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80" exitCode=0 Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.289446 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerDied","Data":"d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.296710 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.318583 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.330157 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.341333 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.354645 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.368444 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.379526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.379572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.379583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.379601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.379614 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.387603 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.399580 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.411182 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.424342 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.436657 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.446944 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.458218 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.475069 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.483187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.483217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.483228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.483245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.483257 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.491535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.507595 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.586177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.586206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.586216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.586230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.586240 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.689081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.689128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.689137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.689155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.689166 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.791639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.791683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.791694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.791711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.791724 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.894628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.894681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.894694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.894717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.894734 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.997077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.997117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.997128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.997146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:54 crc kubenswrapper[4824]: I1124 13:14:54.997158 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:54Z","lastTransitionTime":"2025-11-24T13:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.099545 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.099578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.099585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.099600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.099609 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:55Z","lastTransitionTime":"2025-11-24T13:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.201652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.201686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.201697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.201710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.201722 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:55Z","lastTransitionTime":"2025-11-24T13:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.303144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.303205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.303228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.303273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.303297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:55Z","lastTransitionTime":"2025-11-24T13:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.304681 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0" containerID="bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733" exitCode=0 Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.304742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerDied","Data":"bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.320557 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.334297 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.345992 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.363178 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.382407 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.400264 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.405275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.405304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.405313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.405326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.405335 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:55Z","lastTransitionTime":"2025-11-24T13:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.412534 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.424236 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.435991 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.448884 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.461761 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.476034 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.491709 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.505742 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.507967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.508000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.508014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.508033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.508045 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:55Z","lastTransitionTime":"2025-11-24T13:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.517529 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.611281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.611631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.611640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.611654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.611663 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:55Z","lastTransitionTime":"2025-11-24T13:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.713500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.713554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.713565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.713585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.713597 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:55Z","lastTransitionTime":"2025-11-24T13:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.758376 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.758481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.758559 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.758640 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.758657 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:11.758613015 +0000 UTC m=+53.398152355 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.758669 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.758719 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.758790 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:15:11.75878098 +0000 UTC m=+53.398320290 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.758670 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.758891 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.758912 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.758976 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:15:11.758958005 +0000 UTC m=+53.398497315 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.816293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.816346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.816357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.816388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.816399 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:55Z","lastTransitionTime":"2025-11-24T13:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.859880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.859985 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.860014 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.860081 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.860092 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:15:11.860070781 +0000 UTC m=+53.499610091 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:14:55 crc kubenswrapper[4824]: E1124 13:14:55.860115 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:15:11.860105812 +0000 UTC m=+53.499645122 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.918394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.918451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.918464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.918488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:55 crc kubenswrapper[4824]: I1124 13:14:55.918502 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:55Z","lastTransitionTime":"2025-11-24T13:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.009381 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.009418 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.009381 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:56 crc kubenswrapper[4824]: E1124 13:14:56.009550 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:56 crc kubenswrapper[4824]: E1124 13:14:56.009612 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:56 crc kubenswrapper[4824]: E1124 13:14:56.009684 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.020410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.020454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.020465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.020482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.020495 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.122688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.122737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.122748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.122768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.122781 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.225377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.225415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.225426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.225443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.225454 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.314203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" event={"ID":"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0","Type":"ContainerStarted","Data":"b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.320918 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.321496 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.321679 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.327995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.328067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.328088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.328112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.328126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.332525 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.347032 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.360278 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.364306 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.364691 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.384520 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.396244 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.409462 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.421597 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.430476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.430514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.430524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.430540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.430552 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.436572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.457095 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.473611 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.487263 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.499758 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.511358 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.525146 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.539045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.539096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.539107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.539126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.539137 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.544502 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.565751 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.593149 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.613147 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.628116 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.643416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.643451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.643481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.643499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.643512 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.644432 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.656531 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.667852 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.685149 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.703485 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.721506 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.733975 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.746431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.746470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.746477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.746492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.746502 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.749210 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.760776 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.773188 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.784791 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.848928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.849648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.849743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.849836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.849913 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.953631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.953957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.954058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.954137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:56 crc kubenswrapper[4824]: I1124 13:14:56.954207 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:56Z","lastTransitionTime":"2025-11-24T13:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.060466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.060502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.060512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.060528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.060539 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.256844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.256896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.256909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.256929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.256945 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.259263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.259322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.259331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.259631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.259710 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: E1124 13:14:57.275429 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.279471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.279569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.279579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.279593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.279603 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: E1124 13:14:57.293139 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.303702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.303736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.303745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.303761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.303771 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: E1124 13:14:57.315063 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.319379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.319432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.319447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.319466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.319477 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.323376 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:14:57 crc kubenswrapper[4824]: E1124 13:14:57.332939 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.336276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.336320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.336332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.336347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.336357 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: E1124 13:14:57.347640 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:57 crc kubenswrapper[4824]: E1124 13:14:57.347791 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.360135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.360181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.360190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.360205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.360217 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.463174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.463452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.463544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.463614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.463675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.568021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.568531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.568691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.568898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.569026 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.672345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.672389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.672401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.672421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.672434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.775351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.775444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.775459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.775482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.775501 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.879208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.879246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.879259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.879280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.879291 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.982423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.982477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.982488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.982506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:57 crc kubenswrapper[4824]: I1124 13:14:57.982516 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:57Z","lastTransitionTime":"2025-11-24T13:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.009872 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:14:58 crc kubenswrapper[4824]: E1124 13:14:58.010041 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.010160 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.010237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:14:58 crc kubenswrapper[4824]: E1124 13:14:58.010385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:14:58 crc kubenswrapper[4824]: E1124 13:14:58.010480 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.085934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.085979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.085997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.086018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.086029 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:58Z","lastTransitionTime":"2025-11-24T13:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.189540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.189589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.189599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.189617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.189629 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:58Z","lastTransitionTime":"2025-11-24T13:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.292677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.293180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.293329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.293429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.293550 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:58Z","lastTransitionTime":"2025-11-24T13:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.327141 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.396244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.396590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.396699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.396984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.397200 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:58Z","lastTransitionTime":"2025-11-24T13:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.499768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.500069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.500138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.500208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.500278 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:58Z","lastTransitionTime":"2025-11-24T13:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.602654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.602934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.603016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.603101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.603283 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:58Z","lastTransitionTime":"2025-11-24T13:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.707705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.707785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.707825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.707851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.707877 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:58Z","lastTransitionTime":"2025-11-24T13:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.810928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.810964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.810973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.810989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.810998 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:58Z","lastTransitionTime":"2025-11-24T13:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.913190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.913237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.913247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.913263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:58 crc kubenswrapper[4824]: I1124 13:14:58.913275 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:58Z","lastTransitionTime":"2025-11-24T13:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.015018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.015056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.015069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.015086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.015099 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.022181 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl"] Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.022610 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.024112 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.024263 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.026435 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.042457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.061123 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.071434 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.071548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-249dc\" (UniqueName: \"kubernetes.io/projected/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-kube-api-access-249dc\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.071612 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.071661 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.079758 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.101907 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.118220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.118267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.118282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.118304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.118348 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.124422 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.152854 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.170424 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.172000 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-249dc\" (UniqueName: \"kubernetes.io/projected/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-kube-api-access-249dc\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.172067 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.172098 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.172125 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.172880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.173032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.180480 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.190474 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-249dc\" (UniqueName: \"kubernetes.io/projected/ff0fb887-4240-4e3d-9844-cdfa8ce2d974-kube-api-access-249dc\") pod \"ovnkube-control-plane-749d76644c-ppkzl\" (UID: \"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.193129 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.221847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.222102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.222178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.222253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.222321 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.223171 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.238989 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.256423 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.270212 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.284595 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.303600 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.316099 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.329188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.329226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.329236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.329253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.329264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.329833 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.333999 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.349945 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.366327 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.405834 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.431666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.431714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.431727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.431748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.431761 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.435236 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.450660 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.465184 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.492269 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.511818 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.530140 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.534887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.534924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.534936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.534953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.534964 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.547967 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.564368 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.587913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.601310 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.618436 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.637726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.637769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.637780 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.637798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.637826 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.739834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.739880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.739889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.739910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.739922 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.842395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.842438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.842449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.842467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.842478 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.893466 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.944935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.945011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.945028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.945057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:14:59 crc kubenswrapper[4824]: I1124 13:14:59.945073 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:14:59Z","lastTransitionTime":"2025-11-24T13:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.009720 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.009797 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:00 crc kubenswrapper[4824]: E1124 13:15:00.009891 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.009720 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:00 crc kubenswrapper[4824]: E1124 13:15:00.009980 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:00 crc kubenswrapper[4824]: E1124 13:15:00.010158 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.047662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.047702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.047713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.047730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.047742 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.150965 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.151016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.151026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.151046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.151058 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.254232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.254276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.254286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.254304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.254315 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.340023 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" event={"ID":"ff0fb887-4240-4e3d-9844-cdfa8ce2d974","Type":"ContainerStarted","Data":"8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.340080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" event={"ID":"ff0fb887-4240-4e3d-9844-cdfa8ce2d974","Type":"ContainerStarted","Data":"04ef2f11b667346c6d9365a34a2ac0b48e8b1327b4cfb68453af5f1f5d0b7bfb"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.356826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.356859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.356870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.356884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.356896 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.460178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.460215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.460225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.460241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.460252 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.542447 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-sc887"] Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.543615 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:00 crc kubenswrapper[4824]: E1124 13:15:00.543702 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.563729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.563780 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.563792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.563824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.563837 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.569297 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.587198 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.612882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.629016 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.629087 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42w7z\" (UniqueName: \"kubernetes.io/projected/7e81fafe-6567-47f5-bb87-78fffb25fa9b-kube-api-access-42w7z\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.629137 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.644389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.667288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.667335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.667349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.667372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.667386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.679840 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.695940 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.714710 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.729952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42w7z\" (UniqueName: \"kubernetes.io/projected/7e81fafe-6567-47f5-bb87-78fffb25fa9b-kube-api-access-42w7z\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.730295 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:00 crc kubenswrapper[4824]: E1124 13:15:00.730490 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:00 crc kubenswrapper[4824]: E1124 13:15:00.730613 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs podName:7e81fafe-6567-47f5-bb87-78fffb25fa9b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:01.230583041 +0000 UTC m=+42.870122351 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs") pod "network-metrics-daemon-sc887" (UID: "7e81fafe-6567-47f5-bb87-78fffb25fa9b") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.731705 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.744161 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.750747 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42w7z\" (UniqueName: \"kubernetes.io/projected/7e81fafe-6567-47f5-bb87-78fffb25fa9b-kube-api-access-42w7z\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.755232 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.770027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.770075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.770085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.770100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.770112 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.770855 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.783325 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.793651 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.808654 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.821330 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.835509 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.872657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.872693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.872701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.872716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.872726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.975342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.975396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.975441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.975460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:00 crc kubenswrapper[4824]: I1124 13:15:00.975473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:00Z","lastTransitionTime":"2025-11-24T13:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.078833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.078897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.078914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.078936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.078976 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:01Z","lastTransitionTime":"2025-11-24T13:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.181954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.182025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.182045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.182071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.182086 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:01Z","lastTransitionTime":"2025-11-24T13:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.236375 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:01 crc kubenswrapper[4824]: E1124 13:15:01.236528 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:01 crc kubenswrapper[4824]: E1124 13:15:01.237169 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs podName:7e81fafe-6567-47f5-bb87-78fffb25fa9b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:02.237152479 +0000 UTC m=+43.876691789 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs") pod "network-metrics-daemon-sc887" (UID: "7e81fafe-6567-47f5-bb87-78fffb25fa9b") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.285374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.285420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.285433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.285454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.285472 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:01Z","lastTransitionTime":"2025-11-24T13:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.347630 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/0.log" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.351666 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f" exitCode=1 Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.351769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.353023 4824 scope.go:117] "RemoveContainer" containerID="353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.355680 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" event={"ID":"ff0fb887-4240-4e3d-9844-cdfa8ce2d974","Type":"ContainerStarted","Data":"6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.369974 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.386603 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.397654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.397701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.397713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.397729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.397742 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:01Z","lastTransitionTime":"2025-11-24T13:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.416560 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.433352 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.449586 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.472899 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.490021 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.500920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.500978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.500997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.501023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.501041 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:01Z","lastTransitionTime":"2025-11-24T13:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.504453 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.526940 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"0 6072 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.468988 6072 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.469055 6072 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:15:00.469559 6072 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.470051 6072 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:15:00.470110 6072 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:00.470120 6072 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:00.470184 6072 factory.go:656] Stopping watch factory\\\\nI1124 13:15:00.470206 6072 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:00.470214 6072 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:15:00.470223 6072 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.541206 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.571049 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.584980 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.599066 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.603581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.603613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.603624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.603671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.603688 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:01Z","lastTransitionTime":"2025-11-24T13:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.617569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.633534 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.653280 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.671218 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.686657 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.707062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.707099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.707108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.707125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.707135 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:01Z","lastTransitionTime":"2025-11-24T13:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.714144 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.735605 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.752842 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.768034 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.783890 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.804587 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"0 6072 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.468988 6072 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.469055 6072 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:15:00.469559 6072 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.470051 6072 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:15:00.470110 6072 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:00.470120 6072 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:00.470184 6072 factory.go:656] Stopping watch factory\\\\nI1124 13:15:00.470206 6072 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:00.470214 6072 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:15:00.470223 6072 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.809997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.810039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.810048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.810064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.810076 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:01Z","lastTransitionTime":"2025-11-24T13:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.818969 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.840251 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.852991 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.871968 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.885904 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.898046 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.910467 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.912692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.912752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.912768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.912789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.912802 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:01Z","lastTransitionTime":"2025-11-24T13:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.923324 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.932933 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:01 crc kubenswrapper[4824]: I1124 13:15:01.962300 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.027197 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.027238 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.027293 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:02 crc kubenswrapper[4824]: E1124 13:15:02.027368 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.027369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:02 crc kubenswrapper[4824]: E1124 13:15:02.027475 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:02 crc kubenswrapper[4824]: E1124 13:15:02.027655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:02 crc kubenswrapper[4824]: E1124 13:15:02.027725 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.029390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.029430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.029443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.029463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.029475 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.132151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.132190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.132200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.132218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.132229 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.234648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.234709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.234727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.234756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.234778 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.265060 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:02 crc kubenswrapper[4824]: E1124 13:15:02.265356 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:02 crc kubenswrapper[4824]: E1124 13:15:02.265503 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs podName:7e81fafe-6567-47f5-bb87-78fffb25fa9b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:04.265465744 +0000 UTC m=+45.905005204 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs") pod "network-metrics-daemon-sc887" (UID: "7e81fafe-6567-47f5-bb87-78fffb25fa9b") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.337307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.337350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.337363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.337382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.337396 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.364381 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/0.log" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.369374 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.370428 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.395752 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.413420 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.436379 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.440901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.440971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.440988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.441016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.441035 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.462378 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"0 6072 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.468988 6072 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.469055 6072 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:15:00.469559 6072 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.470051 6072 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:15:00.470110 6072 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:00.470120 6072 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:00.470184 6072 factory.go:656] Stopping watch factory\\\\nI1124 13:15:00.470206 6072 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:00.470214 6072 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:15:00.470223 6072 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.479215 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.506493 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.540312 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.544373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.544426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.544436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.544456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.544468 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.567429 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.581425 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.597187 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.609317 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.620316 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.631274 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.643970 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.647318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.647343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.647354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.647372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.647383 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.659635 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.675236 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.687250 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.775487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.775552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.775565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.775584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.775596 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.878613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.878653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.878665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.878682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.878693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.981014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.981065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.981077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.981097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:02 crc kubenswrapper[4824]: I1124 13:15:02.981113 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:02Z","lastTransitionTime":"2025-11-24T13:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.083022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.083090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.083103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.083131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.083147 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:03Z","lastTransitionTime":"2025-11-24T13:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.186886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.186947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.186967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.186993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.187008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:03Z","lastTransitionTime":"2025-11-24T13:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.290310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.290356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.290368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.290388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.290407 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:03Z","lastTransitionTime":"2025-11-24T13:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.375494 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/1.log" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.376082 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/0.log" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.381099 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8" exitCode=1 Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.381183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.381246 4824 scope.go:117] "RemoveContainer" containerID="353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.382112 4824 scope.go:117] "RemoveContainer" containerID="0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8" Nov 24 13:15:03 crc kubenswrapper[4824]: E1124 13:15:03.382437 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.392845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.392902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.392916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.392939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.392955 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:03Z","lastTransitionTime":"2025-11-24T13:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.400188 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.417451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.432135 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.444729 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.461188 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.479498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.495455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.495494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.495505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.495525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.495537 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:03Z","lastTransitionTime":"2025-11-24T13:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.496355 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.515386 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.531330 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.546438 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.559386 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.575558 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.593905 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://353e7c463d111e18b39faa63a6f9bfb6d18e5f507c0ca934bbe754151606581f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"0 6072 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.468988 6072 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.469055 6072 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:15:00.469559 6072 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:00.470051 6072 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:15:00.470110 6072 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:00.470120 6072 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:00.470184 6072 factory.go:656] Stopping watch factory\\\\nI1124 13:15:00.470206 6072 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:00.470214 6072 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:15:00.470223 6072 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:03Z\\\",\\\"message\\\":\\\"j7 in node crc\\\\nI1124 13:15:03.071906 6271 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-xtlj7 after 0 failed attempt(s)\\\\nI1124 13:15:03.071917 6271 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-xtlj7\\\\nF1124 13:15:03.071921 6271 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:15:03.071932 6271 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071938 6271 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071945 6271 ovn.g\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.598178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.598213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.598224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.598245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.598260 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:03Z","lastTransitionTime":"2025-11-24T13:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.605830 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.625869 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.642414 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.655742 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.702280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.702312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.702321 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.702338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.702349 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:03Z","lastTransitionTime":"2025-11-24T13:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.805260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.805335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.805362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.805397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.805426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:03Z","lastTransitionTime":"2025-11-24T13:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.908734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.908835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.908895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.908928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:03 crc kubenswrapper[4824]: I1124 13:15:03.908955 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:03Z","lastTransitionTime":"2025-11-24T13:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.009750 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.009856 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.009772 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.009993 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:04 crc kubenswrapper[4824]: E1124 13:15:04.010462 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:04 crc kubenswrapper[4824]: E1124 13:15:04.010643 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:04 crc kubenswrapper[4824]: E1124 13:15:04.010729 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:04 crc kubenswrapper[4824]: E1124 13:15:04.010795 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.010878 4824 scope.go:117] "RemoveContainer" containerID="3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.011914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.011954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.011966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.011987 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.012001 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.115717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.115753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.115765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.115827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.115876 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.218418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.218475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.218486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.218509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.218521 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.289905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:04 crc kubenswrapper[4824]: E1124 13:15:04.290090 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:04 crc kubenswrapper[4824]: E1124 13:15:04.290146 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs podName:7e81fafe-6567-47f5-bb87-78fffb25fa9b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:08.290129375 +0000 UTC m=+49.929668685 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs") pod "network-metrics-daemon-sc887" (UID: "7e81fafe-6567-47f5-bb87-78fffb25fa9b") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.321632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.321689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.321699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.321718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.321728 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.387003 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/1.log" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.392760 4824 scope.go:117] "RemoveContainer" containerID="0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8" Nov 24 13:15:04 crc kubenswrapper[4824]: E1124 13:15:04.392994 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.409759 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.424058 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.425258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.425295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.425305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.425323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.425334 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.443045 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.463030 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.487723 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:03Z\\\",\\\"message\\\":\\\"j7 in node crc\\\\nI1124 13:15:03.071906 6271 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-xtlj7 after 0 failed attempt(s)\\\\nI1124 13:15:03.071917 6271 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-xtlj7\\\\nF1124 13:15:03.071921 6271 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:15:03.071932 6271 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071938 6271 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071945 6271 ovn.g\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.498274 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.516053 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.527872 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.528590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.528621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.528633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.528672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.528685 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.542260 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.560829 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.574950 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.587173 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.597428 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.607462 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.617106 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.628214 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.631056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.631088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.631098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.631115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.631134 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.640570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.734132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.734182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.734198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.734217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.734233 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.837661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.837728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.837748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.837775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.837793 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.941096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.941138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.941147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.941164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:04 crc kubenswrapper[4824]: I1124 13:15:04.941174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:04Z","lastTransitionTime":"2025-11-24T13:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.072069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.072230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.072248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.072330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.072346 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.175741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.175830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.175853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.175882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.175901 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.279050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.279126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.279148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.279175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.279192 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.382222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.382269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.382282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.382302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.382318 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.402045 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.403942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.404353 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.420844 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.434153 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.445179 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.455439 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.467972 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.482536 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.484477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.484540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.484558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.484581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.484597 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.498139 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.509023 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.531626 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.544429 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.558932 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.573424 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.587757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.587802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.587831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.587855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.587871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.590421 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.615071 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:03Z\\\",\\\"message\\\":\\\"j7 in node crc\\\\nI1124 13:15:03.071906 6271 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-xtlj7 after 0 failed attempt(s)\\\\nI1124 13:15:03.071917 6271 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-xtlj7\\\\nF1124 13:15:03.071921 6271 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:15:03.071932 6271 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071938 6271 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071945 6271 ovn.g\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.627441 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.642546 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.656138 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.690333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.690381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.690393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.690412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.690426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.792533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.792564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.792572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.792586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.792595 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.895247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.895295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.895306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.895327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.895337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.997926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.998005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.998017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.998037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:05 crc kubenswrapper[4824]: I1124 13:15:05.998050 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:05Z","lastTransitionTime":"2025-11-24T13:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.009754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.009801 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.009754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.009913 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:06 crc kubenswrapper[4824]: E1124 13:15:06.009984 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:06 crc kubenswrapper[4824]: E1124 13:15:06.010310 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:06 crc kubenswrapper[4824]: E1124 13:15:06.010405 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:06 crc kubenswrapper[4824]: E1124 13:15:06.010167 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.100823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.100872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.100882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.100902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.100914 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:06Z","lastTransitionTime":"2025-11-24T13:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.204783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.204860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.204872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.204892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.204906 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:06Z","lastTransitionTime":"2025-11-24T13:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.307992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.308040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.308050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.308068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.308083 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:06Z","lastTransitionTime":"2025-11-24T13:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.410415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.410449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.410459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.410475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.410484 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:06Z","lastTransitionTime":"2025-11-24T13:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.523325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.523423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.523450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.523482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.523504 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:06Z","lastTransitionTime":"2025-11-24T13:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.626130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.626173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.626184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.626202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.626212 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:06Z","lastTransitionTime":"2025-11-24T13:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.729058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.729118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.729135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.729166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.729184 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:06Z","lastTransitionTime":"2025-11-24T13:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.832522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.832595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.832614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.832645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.832665 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:06Z","lastTransitionTime":"2025-11-24T13:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.936960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.937034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.937058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.937087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:06 crc kubenswrapper[4824]: I1124 13:15:06.937108 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:06Z","lastTransitionTime":"2025-11-24T13:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.040854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.040902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.040912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.040928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.040939 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.144416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.144490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.144507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.144535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.144550 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.248010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.248083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.248103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.248134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.248157 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.351375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.351416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.351427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.351448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.351459 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.454623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.454680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.454695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.454742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.454758 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.496186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.496236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.496252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.496279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.496296 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: E1124 13:15:07.509073 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.512948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.512983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.512997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.513015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.513029 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: E1124 13:15:07.526589 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.531197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.531253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.531263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.531288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.531299 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: E1124 13:15:07.547068 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.551495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.551555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.551569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.551591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.551605 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: E1124 13:15:07.565943 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.570735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.570800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.570830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.570852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.570866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: E1124 13:15:07.589110 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:07Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:07 crc kubenswrapper[4824]: E1124 13:15:07.589362 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.591003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.591038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.591052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.591072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.591085 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.694378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.694412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.694421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.694445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.694456 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.797055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.797128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.797143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.797163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.797176 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.899994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.900031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.900039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.900055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:07 crc kubenswrapper[4824]: I1124 13:15:07.900064 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:07Z","lastTransitionTime":"2025-11-24T13:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.002398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.002441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.002461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.002483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.002496 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.009738 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.009773 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.009782 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.009819 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:08 crc kubenswrapper[4824]: E1124 13:15:08.009904 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:08 crc kubenswrapper[4824]: E1124 13:15:08.009993 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:08 crc kubenswrapper[4824]: E1124 13:15:08.010074 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:08 crc kubenswrapper[4824]: E1124 13:15:08.010128 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.105280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.105330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.105343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.105362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.105374 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.208654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.208702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.208713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.208734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.208747 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.311849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.311905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.311914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.311941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.311952 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.330934 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:08 crc kubenswrapper[4824]: E1124 13:15:08.331117 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:08 crc kubenswrapper[4824]: E1124 13:15:08.331220 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs podName:7e81fafe-6567-47f5-bb87-78fffb25fa9b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:16.331200001 +0000 UTC m=+57.970739311 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs") pod "network-metrics-daemon-sc887" (UID: "7e81fafe-6567-47f5-bb87-78fffb25fa9b") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.414927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.415441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.415536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.415631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.415711 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.520019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.520064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.520075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.520094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.520106 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.622971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.623030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.623044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.623068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.623086 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.726064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.726113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.726126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.726148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.726160 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.829685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.829738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.829750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.829770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.829782 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.933001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.933050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.933060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.933081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:08 crc kubenswrapper[4824]: I1124 13:15:08.933098 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:08Z","lastTransitionTime":"2025-11-24T13:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.025011 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.037128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.037201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.037210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.037228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.037240 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.039236 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.053643 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.065630 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.075730 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.090551 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.105216 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.118620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.132576 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.141011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.141067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.141076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.141098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.141439 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.158871 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.175028 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.189252 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.202568 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.218000 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.236937 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:03Z\\\",\\\"message\\\":\\\"j7 in node crc\\\\nI1124 13:15:03.071906 6271 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-xtlj7 after 0 failed attempt(s)\\\\nI1124 13:15:03.071917 6271 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-xtlj7\\\\nF1124 13:15:03.071921 6271 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:15:03.071932 6271 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071938 6271 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071945 6271 ovn.g\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.243944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.243988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.243997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.244019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.244031 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.250484 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.261669 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:09Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.347235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.347289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.347304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.347325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.347339 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.450195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.450234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.450252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.450272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.450285 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.553573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.553632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.553642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.553661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.553675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.657504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.657562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.657575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.657596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.657611 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.760318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.760365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.760375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.760392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.760404 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.863258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.863304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.863312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.863329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.863342 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.966389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.966447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.966458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.966478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:09 crc kubenswrapper[4824]: I1124 13:15:09.966491 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:09Z","lastTransitionTime":"2025-11-24T13:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.009909 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.010024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.010087 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.010086 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:10 crc kubenswrapper[4824]: E1124 13:15:10.010228 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:10 crc kubenswrapper[4824]: E1124 13:15:10.010399 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:10 crc kubenswrapper[4824]: E1124 13:15:10.010481 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:10 crc kubenswrapper[4824]: E1124 13:15:10.010536 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.069082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.069131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.069145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.069167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.069180 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:10Z","lastTransitionTime":"2025-11-24T13:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.172726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.172770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.172782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.172828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.172841 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:10Z","lastTransitionTime":"2025-11-24T13:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.276244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.276287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.276315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.276336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.276349 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:10Z","lastTransitionTime":"2025-11-24T13:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.379685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.379738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.379762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.379787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.379827 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:10Z","lastTransitionTime":"2025-11-24T13:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.482876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.482920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.482929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.482946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.482967 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:10Z","lastTransitionTime":"2025-11-24T13:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.586411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.586467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.586482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.586509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.586525 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:10Z","lastTransitionTime":"2025-11-24T13:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.690764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.690886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.690905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.690938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.690957 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:10Z","lastTransitionTime":"2025-11-24T13:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.795355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.795417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.795431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.795457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.795473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:10Z","lastTransitionTime":"2025-11-24T13:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.901717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.901817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.901830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.902045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:10 crc kubenswrapper[4824]: I1124 13:15:10.902055 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:10Z","lastTransitionTime":"2025-11-24T13:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.025195 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.025371 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.028431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.028505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.028522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.028552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.028574 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.132567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.132615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.132627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.132672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.132686 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.235436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.235487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.235499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.235529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.235543 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.338440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.338537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.338560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.338592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.338618 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.441878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.441931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.441946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.441968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.441985 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.545140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.545188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.545198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.545217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.545232 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.648068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.648129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.648139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.648165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.648177 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.751136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.751200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.751220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.751248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.751269 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.776083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.776425 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:43.776372731 +0000 UTC m=+85.415912091 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.776560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.776744 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.776860 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.777197 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.777227 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.776967 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.777314 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:15:43.777289086 +0000 UTC m=+85.416828436 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.777320 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.777355 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.777434 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:15:43.777409089 +0000 UTC m=+85.416948479 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.854581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.854635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.854646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.854665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.854677 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.877987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.878138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.878297 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.878405 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:15:43.878380341 +0000 UTC m=+85.517919681 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.878555 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:15:11 crc kubenswrapper[4824]: E1124 13:15:11.878670 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:15:43.878643868 +0000 UTC m=+85.518183358 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.958094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.958488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.958581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.958679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:11 crc kubenswrapper[4824]: I1124 13:15:11.958787 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:11Z","lastTransitionTime":"2025-11-24T13:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.010000 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.010080 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:12 crc kubenswrapper[4824]: E1124 13:15:12.010204 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:12 crc kubenswrapper[4824]: E1124 13:15:12.010333 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.010929 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:12 crc kubenswrapper[4824]: E1124 13:15:12.011269 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.061355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.061438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.061458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.061485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.061508 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.170418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.170497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.170509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.170543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.170559 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.273241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.273326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.273337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.273360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.273373 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.376492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.376538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.376554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.376574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.376585 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.479370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.479418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.479430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.479451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.479467 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.582633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.582691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.582701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.582720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.582732 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.686556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.686629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.686651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.686680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.686695 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.789641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.789709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.789734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.789762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.789785 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.892412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.892455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.892465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.892484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.892496 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.996513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.996565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.996581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.996611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:12 crc kubenswrapper[4824]: I1124 13:15:12.996629 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:12Z","lastTransitionTime":"2025-11-24T13:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.010021 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:13 crc kubenswrapper[4824]: E1124 13:15:13.010243 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.099489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.100103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.100323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.100505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.100653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:13Z","lastTransitionTime":"2025-11-24T13:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.203861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.203929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.203944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.203969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.203984 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:13Z","lastTransitionTime":"2025-11-24T13:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.307628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.307683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.307696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.307736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.307748 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:13Z","lastTransitionTime":"2025-11-24T13:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.411531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.411578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.411589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.411613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.411626 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:13Z","lastTransitionTime":"2025-11-24T13:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.515427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.515485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.515498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.515524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.515538 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:13Z","lastTransitionTime":"2025-11-24T13:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.618989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.619041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.619056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.619087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.619108 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:13Z","lastTransitionTime":"2025-11-24T13:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.723273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.723353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.723376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.723407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.723431 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:13Z","lastTransitionTime":"2025-11-24T13:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.826898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.826990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.827011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.827041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.827061 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:13Z","lastTransitionTime":"2025-11-24T13:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.930256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.930322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.930359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.930384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:13 crc kubenswrapper[4824]: I1124 13:15:13.930401 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:13Z","lastTransitionTime":"2025-11-24T13:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.009974 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.010053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:14 crc kubenswrapper[4824]: E1124 13:15:14.010188 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.010087 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:14 crc kubenswrapper[4824]: E1124 13:15:14.010349 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:14 crc kubenswrapper[4824]: E1124 13:15:14.010457 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.034755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.034854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.034876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.034911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.034932 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.139486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.139559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.139579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.139610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.139631 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.210787 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.222854 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.233752 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.242618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.242668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.242689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.242711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.242726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.253260 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.268545 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.281791 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.296936 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.313260 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.335513 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.346256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.346305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.346317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.346375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.346390 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.356573 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.371688 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.383791 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.398595 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.420906 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:03Z\\\",\\\"message\\\":\\\"j7 in node crc\\\\nI1124 13:15:03.071906 6271 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-xtlj7 after 0 failed attempt(s)\\\\nI1124 13:15:03.071917 6271 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-xtlj7\\\\nF1124 13:15:03.071921 6271 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:15:03.071932 6271 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071938 6271 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071945 6271 ovn.g\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.432621 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.448992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.449053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.449069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.449090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.449153 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.452258 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.465883 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.478579 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.490685 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.551952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.552023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.552037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.552078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.552093 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.654777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.654846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.654858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.654877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.654890 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.757709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.757768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.757801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.757833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.757842 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.860047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.860106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.860118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.860142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.860156 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.963757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.963850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.963863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.963883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:14 crc kubenswrapper[4824]: I1124 13:15:14.963897 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:14Z","lastTransitionTime":"2025-11-24T13:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.009834 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:15 crc kubenswrapper[4824]: E1124 13:15:15.009994 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.066964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.067004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.067017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.067037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.067050 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.170436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.170499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.170521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.170555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.170578 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.274529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.274604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.274627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.274661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.274685 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.377516 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.377554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.377563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.377580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.377589 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.481317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.481371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.481384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.481408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.481426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.584942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.585023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.585041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.585072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.585090 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.688137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.688186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.688195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.688212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.688223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.791580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.791636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.791646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.791663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.791674 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.894996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.895068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.895097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.895128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.895151 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.998510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.998558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.998569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.998588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:15 crc kubenswrapper[4824]: I1124 13:15:15.998599 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:15Z","lastTransitionTime":"2025-11-24T13:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.009150 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.009209 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:16 crc kubenswrapper[4824]: E1124 13:15:16.009266 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.009226 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:16 crc kubenswrapper[4824]: E1124 13:15:16.009482 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:16 crc kubenswrapper[4824]: E1124 13:15:16.009608 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.101791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.101884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.101900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.101925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.101942 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:16Z","lastTransitionTime":"2025-11-24T13:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.210672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.210731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.210750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.210782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.210804 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:16Z","lastTransitionTime":"2025-11-24T13:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.313991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.314073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.314089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.314112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.314125 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:16Z","lastTransitionTime":"2025-11-24T13:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.417248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.417317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.417338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.417363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.417382 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:16Z","lastTransitionTime":"2025-11-24T13:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.432289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:16 crc kubenswrapper[4824]: E1124 13:15:16.432487 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:16 crc kubenswrapper[4824]: E1124 13:15:16.432561 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs podName:7e81fafe-6567-47f5-bb87-78fffb25fa9b nodeName:}" failed. No retries permitted until 2025-11-24 13:15:32.432544394 +0000 UTC m=+74.072083704 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs") pod "network-metrics-daemon-sc887" (UID: "7e81fafe-6567-47f5-bb87-78fffb25fa9b") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.520537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.520602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.520613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.520632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.520651 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:16Z","lastTransitionTime":"2025-11-24T13:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.623875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.623924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.623935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.623953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.623966 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:16Z","lastTransitionTime":"2025-11-24T13:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.726881 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.726930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.726940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.726960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.726972 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:16Z","lastTransitionTime":"2025-11-24T13:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.830417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.830517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.830536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.830565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.830589 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:16Z","lastTransitionTime":"2025-11-24T13:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.934923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.935002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.935021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.935056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:16 crc kubenswrapper[4824]: I1124 13:15:16.935079 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:16Z","lastTransitionTime":"2025-11-24T13:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.010062 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:17 crc kubenswrapper[4824]: E1124 13:15:17.010262 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.037418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.037451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.037465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.037489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.037504 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.099645 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.129888 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.142483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.142885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.143120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.143329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.143523 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.159874 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.182882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.223323 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.240053 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.245903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.246126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.246270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.246417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.246529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.255620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.270687 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.294165 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.314340 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:03Z\\\",\\\"message\\\":\\\"j7 in node crc\\\\nI1124 13:15:03.071906 6271 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-xtlj7 after 0 failed attempt(s)\\\\nI1124 13:15:03.071917 6271 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-xtlj7\\\\nF1124 13:15:03.071921 6271 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:15:03.071932 6271 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071938 6271 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071945 6271 ovn.g\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.332882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.348297 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.349164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.349214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.349240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.349262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.349277 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.362157 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.374793 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.390388 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.407643 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.420271 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.431570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.444827 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.452310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.452369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.452386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.452407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.452422 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.555428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.555823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.555922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.556076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.556178 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.658927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.658973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.658983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.659002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.659011 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.731578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.731966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.732309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.732395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.732461 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: E1124 13:15:17.748316 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.753437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.753489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.753503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.753525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.753539 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: E1124 13:15:17.776693 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.782491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.782552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.782566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.782591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.782607 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: E1124 13:15:17.798165 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.803465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.803510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.803525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.803548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.803564 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: E1124 13:15:17.818035 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.823248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.823309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.823323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.823347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.823360 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: E1124 13:15:17.836676 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:17 crc kubenswrapper[4824]: E1124 13:15:17.836872 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.839645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.839784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.839824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.839847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.839862 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.943022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.943059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.943256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.943275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:17 crc kubenswrapper[4824]: I1124 13:15:17.943285 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:17Z","lastTransitionTime":"2025-11-24T13:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.009862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.009917 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.009938 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:18 crc kubenswrapper[4824]: E1124 13:15:18.010311 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:18 crc kubenswrapper[4824]: E1124 13:15:18.010490 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:18 crc kubenswrapper[4824]: E1124 13:15:18.010550 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.010753 4824 scope.go:117] "RemoveContainer" containerID="0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.046420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.046465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.046477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.046497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.046511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.149396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.149955 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.149971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.149995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.150007 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.252835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.252875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.252886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.252904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.252916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.356730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.356766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.356779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.356796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.356827 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.453836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/1.log" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.457921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.459090 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.464171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.464235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.464575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.464610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.464630 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.483289 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.501215 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.520222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.536174 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.553156 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.568121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.568173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.568183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.568208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.568221 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.568536 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.587154 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.610418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:03Z\\\",\\\"message\\\":\\\"j7 in node crc\\\\nI1124 13:15:03.071906 6271 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-xtlj7 after 0 failed attempt(s)\\\\nI1124 13:15:03.071917 6271 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-xtlj7\\\\nF1124 13:15:03.071921 6271 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:15:03.071932 6271 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071938 6271 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071945 6271 ovn.g\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.626782 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.652286 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.671421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.671469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.671479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.671500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.671510 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.672230 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.697506 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.714187 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.738152 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.755713 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.774085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.774149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.774168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.774196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.774216 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.791961 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.804604 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.818482 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.876910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.876964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.876974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.876992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.877004 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.981025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.981087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.981102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.981127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:18 crc kubenswrapper[4824]: I1124 13:15:18.981143 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:18Z","lastTransitionTime":"2025-11-24T13:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.010226 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:19 crc kubenswrapper[4824]: E1124 13:15:19.010385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.028901 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.041912 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.054183 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.064641 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.084918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.084961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.084973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.084994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.085008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:19Z","lastTransitionTime":"2025-11-24T13:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.089479 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.106223 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.124505 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.144778 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.163388 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.188370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.188430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.188441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.188463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.188474 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:19Z","lastTransitionTime":"2025-11-24T13:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.189109 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:03Z\\\",\\\"message\\\":\\\"j7 in node crc\\\\nI1124 13:15:03.071906 6271 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-xtlj7 after 0 failed attempt(s)\\\\nI1124 13:15:03.071917 6271 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-xtlj7\\\\nF1124 13:15:03.071921 6271 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:15:03.071932 6271 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071938 6271 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071945 6271 ovn.g\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.203545 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.218150 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.235055 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.248557 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.260239 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.272056 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.282946 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.290729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.290779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.290789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.290821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.290836 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:19Z","lastTransitionTime":"2025-11-24T13:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.293666 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.393392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.393462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.393475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.393494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.393508 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:19Z","lastTransitionTime":"2025-11-24T13:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.464095 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/2.log" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.464664 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/1.log" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.468404 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62" exitCode=1 Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.468474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.468553 4824 scope.go:117] "RemoveContainer" containerID="0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.469211 4824 scope.go:117] "RemoveContainer" containerID="1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62" Nov 24 13:15:19 crc kubenswrapper[4824]: E1124 13:15:19.469432 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.490210 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.496404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.496436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.496445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.496462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.496473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:19Z","lastTransitionTime":"2025-11-24T13:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.505251 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.522026 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.537231 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.550204 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.565664 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.581379 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.598607 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.599799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.599907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.599927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.599973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.599990 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:19Z","lastTransitionTime":"2025-11-24T13:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.612222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.627278 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.641615 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.669221 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.683287 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.698193 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.703332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.703386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.703399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.703419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.703430 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:19Z","lastTransitionTime":"2025-11-24T13:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.718535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e0bd376efc11171a99b89df70f032781c522d5b51db80e50e2c93be0114d9e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:03Z\\\",\\\"message\\\":\\\"j7 in node crc\\\\nI1124 13:15:03.071906 6271 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-xtlj7 after 0 failed attempt(s)\\\\nI1124 13:15:03.071917 6271 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-xtlj7\\\\nF1124 13:15:03.071921 6271 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:03Z is after 2025-08-24T17:21:41Z]\\\\nI1124 13:15:03.071932 6271 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071938 6271 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1124 13:15:03.071945 6271 ovn.g\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:19Z\\\",\\\"message\\\":\\\"1.EgressIP event handler 8\\\\nI1124 13:15:19.170559 6461 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:19.170570 6461 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:19.170602 6461 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:15:19.170613 6461 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:19.170619 6461 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:15:19.170626 6461 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:19.170637 6461 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:19.170645 6461 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:15:19.170661 6461 factory.go:656] Stopping watch factory\\\\nI1124 13:15:19.170680 6461 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:15:19.170688 6461 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:15:19.170837 6461 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1124 13:15:19.170920 6461 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:15:19.170958 6461 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:19.170980 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:19.171073 6461 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.736310 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.757583 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.772018 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.806084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.806214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.806233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.806252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.806265 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:19Z","lastTransitionTime":"2025-11-24T13:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.909765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.909842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.909854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.909879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:19 crc kubenswrapper[4824]: I1124 13:15:19.909892 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:19Z","lastTransitionTime":"2025-11-24T13:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.010105 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.010203 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.010290 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:20 crc kubenswrapper[4824]: E1124 13:15:20.010848 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:20 crc kubenswrapper[4824]: E1124 13:15:20.011041 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:20 crc kubenswrapper[4824]: E1124 13:15:20.011132 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.011927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.011964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.011979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.011997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.012010 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.115591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.115636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.115661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.115681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.115693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.220086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.220244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.220269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.220302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.220327 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.323681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.323748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.323761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.323784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.323797 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.426746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.426831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.426846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.426869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.426886 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.474507 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/2.log" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.478113 4824 scope.go:117] "RemoveContainer" containerID="1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62" Nov 24 13:15:20 crc kubenswrapper[4824]: E1124 13:15:20.478408 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.491407 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.502882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.512308 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.524745 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.529493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.529528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.529538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.529571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.529587 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.538408 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.552567 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.568119 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.586448 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.601135 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.616110 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.631695 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.632713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.632767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.632778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.632828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.632842 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.655580 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:19Z\\\",\\\"message\\\":\\\"1.EgressIP event handler 8\\\\nI1124 13:15:19.170559 6461 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:19.170570 6461 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:19.170602 6461 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:15:19.170613 6461 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:19.170619 6461 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:15:19.170626 6461 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:19.170637 6461 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:19.170645 6461 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:15:19.170661 6461 factory.go:656] Stopping watch factory\\\\nI1124 13:15:19.170680 6461 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:15:19.170688 6461 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:15:19.170837 6461 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1124 13:15:19.170920 6461 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:15:19.170958 6461 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:19.170980 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:19.171073 6461 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.671112 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.696715 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.714281 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.731175 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.735400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.735442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.735453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.735469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.735479 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.746175 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.761393 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.838004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.838091 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.838101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.838116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.838127 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.940522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.940575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.940600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.940624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:20 crc kubenswrapper[4824]: I1124 13:15:20.940635 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:20Z","lastTransitionTime":"2025-11-24T13:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.010157 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:21 crc kubenswrapper[4824]: E1124 13:15:21.010350 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.043576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.043634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.043654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.043684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.043718 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.146600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.146650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.146665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.146684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.146697 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.251165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.251420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.251494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.251610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.251698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.353743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.353973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.354119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.354199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.354257 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.456127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.456168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.456178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.456194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.456206 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.558851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.559103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.559178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.559271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.559332 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.662700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.663020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.663093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.663161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.663223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.766379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.766444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.766463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.766495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.766509 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.869190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.869237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.869246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.869262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.869272 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.972191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.972233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.972246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.972264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:21 crc kubenswrapper[4824]: I1124 13:15:21.972274 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:21Z","lastTransitionTime":"2025-11-24T13:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.009374 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.009515 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:22 crc kubenswrapper[4824]: E1124 13:15:22.009561 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.009375 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:22 crc kubenswrapper[4824]: E1124 13:15:22.009783 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:22 crc kubenswrapper[4824]: E1124 13:15:22.009883 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.074630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.074689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.074707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.074732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.074748 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:22Z","lastTransitionTime":"2025-11-24T13:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.178027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.178077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.178087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.178105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.178116 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:22Z","lastTransitionTime":"2025-11-24T13:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.281330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.281384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.281396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.281416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.281431 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:22Z","lastTransitionTime":"2025-11-24T13:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.384624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.384677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.384692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.384717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.384742 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:22Z","lastTransitionTime":"2025-11-24T13:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.487327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.487365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.487374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.487407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.487416 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:22Z","lastTransitionTime":"2025-11-24T13:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.590399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.590485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.590500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.590526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.590539 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:22Z","lastTransitionTime":"2025-11-24T13:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.693772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.693844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.693860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.693887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.693911 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:22Z","lastTransitionTime":"2025-11-24T13:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.797319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.797363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.797372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.797390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.797401 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:22Z","lastTransitionTime":"2025-11-24T13:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.899893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.899932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.899941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.899958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:22 crc kubenswrapper[4824]: I1124 13:15:22.899969 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:22Z","lastTransitionTime":"2025-11-24T13:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.002880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.002956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.002973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.002996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.003008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.009260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:23 crc kubenswrapper[4824]: E1124 13:15:23.009400 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.106328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.106376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.106387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.106404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.106415 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.210063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.210123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.210135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.210165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.210183 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.312701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.312739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.312747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.312765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.312775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.415855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.415917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.415940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.415967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.415985 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.518919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.518969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.518980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.519000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.519013 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.622295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.622346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.622358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.622382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.622393 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.725468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.725516 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.725526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.725545 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.725557 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.833789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.833849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.833860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.833879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.833894 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.936569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.936626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.936636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.936653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:23 crc kubenswrapper[4824]: I1124 13:15:23.936664 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:23Z","lastTransitionTime":"2025-11-24T13:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.009584 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.009669 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:24 crc kubenswrapper[4824]: E1124 13:15:24.009774 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:24 crc kubenswrapper[4824]: E1124 13:15:24.009947 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.010240 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:24 crc kubenswrapper[4824]: E1124 13:15:24.010380 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.040005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.040067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.040082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.040102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.040117 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.144519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.144573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.144584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.144611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.144624 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.247991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.248041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.248052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.248075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.248087 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.352317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.352422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.352918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.352981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.353012 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.456881 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.456925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.456936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.456956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.456971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.560937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.561034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.561059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.561097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.561120 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.664154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.664197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.664208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.664227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.664238 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.767252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.767307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.767318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.767339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.767348 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.870873 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.870913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.870922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.870938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.870948 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.973790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.973870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.973885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.973912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:24 crc kubenswrapper[4824]: I1124 13:15:24.973929 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:24Z","lastTransitionTime":"2025-11-24T13:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.009823 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:25 crc kubenswrapper[4824]: E1124 13:15:25.010218 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.030191 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.076848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.076905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.076954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.076977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.076991 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:25Z","lastTransitionTime":"2025-11-24T13:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.179683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.179727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.179737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.179754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.179766 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:25Z","lastTransitionTime":"2025-11-24T13:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.282368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.282414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.282422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.282439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.282449 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:25Z","lastTransitionTime":"2025-11-24T13:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.384040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.384087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.384104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.384123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.384135 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:25Z","lastTransitionTime":"2025-11-24T13:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.487281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.487328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.487337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.487359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.487368 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:25Z","lastTransitionTime":"2025-11-24T13:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.589635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.589948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.590082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.590241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.590425 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:25Z","lastTransitionTime":"2025-11-24T13:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.693156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.693559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.693649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.693739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.693853 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:25Z","lastTransitionTime":"2025-11-24T13:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.797357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.797902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.798124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.798271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.798409 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:25Z","lastTransitionTime":"2025-11-24T13:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.901571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.902848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.902934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.903046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:25 crc kubenswrapper[4824]: I1124 13:15:25.903115 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:25Z","lastTransitionTime":"2025-11-24T13:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.006231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.006282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.006294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.006313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.006329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.009507 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.009539 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:26 crc kubenswrapper[4824]: E1124 13:15:26.009631 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:26 crc kubenswrapper[4824]: E1124 13:15:26.009747 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.010133 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:26 crc kubenswrapper[4824]: E1124 13:15:26.010357 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.109475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.109516 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.109525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.109543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.109553 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.211988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.212275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.212341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.212411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.212509 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.315742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.315798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.315839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.315861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.315894 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.418296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.418722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.418834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.418936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.418998 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.521767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.521924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.521943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.521966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.521978 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.624603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.624671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.624684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.624708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.624722 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.727261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.727306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.727317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.727334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.727344 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.829637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.829671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.829682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.829698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.829708 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.932285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.932344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.932358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.932383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:26 crc kubenswrapper[4824]: I1124 13:15:26.932398 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:26Z","lastTransitionTime":"2025-11-24T13:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.009292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:27 crc kubenswrapper[4824]: E1124 13:15:27.009530 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.034888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.034934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.034946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.034965 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.034977 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.138154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.138196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.138206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.138223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.138235 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.241365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.241431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.241448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.241475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.241489 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.344568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.344602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.344614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.344631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.344641 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.447484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.447527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.447537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.447557 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.447568 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.550450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.550502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.550513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.550536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.550550 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.653287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.653396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.653407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.653427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.653441 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.756376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.756425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.756435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.756454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.756469 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.859781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.859878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.860248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.860365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.860382 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.963910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.963964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.963976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.963997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:27 crc kubenswrapper[4824]: I1124 13:15:27.964013 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:27Z","lastTransitionTime":"2025-11-24T13:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.009716 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.009776 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.009838 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:28 crc kubenswrapper[4824]: E1124 13:15:28.009947 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:28 crc kubenswrapper[4824]: E1124 13:15:28.010214 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:28 crc kubenswrapper[4824]: E1124 13:15:28.010305 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.067921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.067996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.068009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.068029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.068040 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.171208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.171256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.171266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.171287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.171299 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.233262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.233740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.233862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.233942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.233999 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: E1124 13:15:28.250840 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.255621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.255663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.255674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.255711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.255726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: E1124 13:15:28.269599 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.275781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.275872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.275885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.275909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.275928 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: E1124 13:15:28.289889 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.296127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.296170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.296189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.296215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.296233 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: E1124 13:15:28.311181 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.316583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.316615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.316625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.316652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.316663 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: E1124 13:15:28.329587 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:28 crc kubenswrapper[4824]: E1124 13:15:28.329711 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.331908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.331977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.331991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.332017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.332031 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.435440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.435484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.435499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.435523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.435536 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.538311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.538376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.538388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.538416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.538430 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.641089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.641128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.641137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.641153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.641164 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.749868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.749961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.749976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.749998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.750013 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.854275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.854350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.854368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.854395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.854414 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.957328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.957378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.957390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.957414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:28 crc kubenswrapper[4824]: I1124 13:15:28.957427 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:28Z","lastTransitionTime":"2025-11-24T13:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.009218 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:29 crc kubenswrapper[4824]: E1124 13:15:29.009395 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.030127 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.043515 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.059392 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.059894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.059929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.059938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.059954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.059968 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.074294 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.093171 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.116147 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:19Z\\\",\\\"message\\\":\\\"1.EgressIP event handler 8\\\\nI1124 13:15:19.170559 6461 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:19.170570 6461 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:19.170602 6461 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:15:19.170613 6461 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:19.170619 6461 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:15:19.170626 6461 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:19.170637 6461 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:19.170645 6461 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:15:19.170661 6461 factory.go:656] Stopping watch factory\\\\nI1124 13:15:19.170680 6461 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:15:19.170688 6461 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:15:19.170837 6461 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1124 13:15:19.170920 6461 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:15:19.170958 6461 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:19.170980 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:19.171073 6461 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.127631 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.143742 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.161472 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.162746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.162797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.162824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.162844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.162855 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.177005 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c5aeab-efe8-4aac-a669-339d07505417\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.193948 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.211004 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.225232 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.239331 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.252575 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.265418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.265460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.265472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.265492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.265503 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.270451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.288497 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.306128 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.320603 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:29Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.368145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.368192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.368205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.368224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.368238 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.470617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.470659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.470670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.470687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.470699 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.573271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.573331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.573344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.573364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.573373 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.675860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.675914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.675924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.675942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.676272 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.781046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.781103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.781121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.781157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.781171 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.884480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.884537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.884547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.884566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.884577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.987550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.987625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.987640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.987682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:29 crc kubenswrapper[4824]: I1124 13:15:29.987698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:29Z","lastTransitionTime":"2025-11-24T13:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.009887 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:30 crc kubenswrapper[4824]: E1124 13:15:30.010083 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.010146 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:30 crc kubenswrapper[4824]: E1124 13:15:30.010240 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.010887 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:30 crc kubenswrapper[4824]: E1124 13:15:30.011018 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.090522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.090581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.090592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.090636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.090650 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:30Z","lastTransitionTime":"2025-11-24T13:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.192867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.192923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.192935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.192956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.192966 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:30Z","lastTransitionTime":"2025-11-24T13:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.295029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.295079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.295089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.295105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.295114 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:30Z","lastTransitionTime":"2025-11-24T13:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.397968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.398010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.398021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.398038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.398051 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:30Z","lastTransitionTime":"2025-11-24T13:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.500565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.500626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.500639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.500662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.500676 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:30Z","lastTransitionTime":"2025-11-24T13:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.603486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.603535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.603548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.603565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.603577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:30Z","lastTransitionTime":"2025-11-24T13:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.706920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.707011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.707025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.707047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.707060 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:30Z","lastTransitionTime":"2025-11-24T13:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.810206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.810275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.810289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.810311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.810325 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:30Z","lastTransitionTime":"2025-11-24T13:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.912898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.912958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.912971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.912989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:30 crc kubenswrapper[4824]: I1124 13:15:30.913056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:30Z","lastTransitionTime":"2025-11-24T13:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.009626 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:31 crc kubenswrapper[4824]: E1124 13:15:31.009798 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.015791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.015852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.015866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.015885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.015897 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.118978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.119034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.119072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.119101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.119112 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.222672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.222725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.222743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.222763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.222775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.325854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.325894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.325905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.325922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.325932 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.428785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.428870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.428892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.428919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.428937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.531952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.531991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.532001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.532017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.532028 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.634985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.635020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.635028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.635043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.635052 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.741415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.741455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.741466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.741484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.741494 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.843763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.843826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.843838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.843854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.843865 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.946065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.946096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.946104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.946118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:31 crc kubenswrapper[4824]: I1124 13:15:31.946127 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:31Z","lastTransitionTime":"2025-11-24T13:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.009843 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.009843 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:32 crc kubenswrapper[4824]: E1124 13:15:32.009994 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:32 crc kubenswrapper[4824]: E1124 13:15:32.010066 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.009852 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:32 crc kubenswrapper[4824]: E1124 13:15:32.010151 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.048578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.048612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.048620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.048636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.048645 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.155774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.155862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.155875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.155893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.155904 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.258776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.258861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.258872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.258890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.258901 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.361595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.361640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.361651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.361668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.361682 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.464948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.465004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.465019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.465040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.465052 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.508933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:32 crc kubenswrapper[4824]: E1124 13:15:32.509235 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:32 crc kubenswrapper[4824]: E1124 13:15:32.509355 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs podName:7e81fafe-6567-47f5-bb87-78fffb25fa9b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:04.509334334 +0000 UTC m=+106.148873644 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs") pod "network-metrics-daemon-sc887" (UID: "7e81fafe-6567-47f5-bb87-78fffb25fa9b") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.568280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.568326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.568337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.568355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.568366 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.673044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.673089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.673101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.673124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.673134 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.776229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.776289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.776307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.776334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.776347 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.879082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.879144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.879153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.879173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.879184 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.982045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.982094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.982105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.982122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:32 crc kubenswrapper[4824]: I1124 13:15:32.982132 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:32Z","lastTransitionTime":"2025-11-24T13:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.010073 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:33 crc kubenswrapper[4824]: E1124 13:15:33.010265 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.084622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.084671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.084682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.084700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.084711 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:33Z","lastTransitionTime":"2025-11-24T13:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.187498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.187546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.187555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.187575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.187587 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:33Z","lastTransitionTime":"2025-11-24T13:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.290457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.290509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.290517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.290535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.290546 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:33Z","lastTransitionTime":"2025-11-24T13:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.392470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.392519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.392530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.392546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.392556 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:33Z","lastTransitionTime":"2025-11-24T13:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.495034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.495075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.495085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.495100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.495110 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:33Z","lastTransitionTime":"2025-11-24T13:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.598441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.598498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.598515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.598538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.598551 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:33Z","lastTransitionTime":"2025-11-24T13:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.701638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.701696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.701709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.701729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.701741 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:33Z","lastTransitionTime":"2025-11-24T13:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.804498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.804573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.804587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.804606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.804618 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:33Z","lastTransitionTime":"2025-11-24T13:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.907831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.907878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.907889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.907905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:33 crc kubenswrapper[4824]: I1124 13:15:33.907916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:33Z","lastTransitionTime":"2025-11-24T13:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.009131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.009212 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.009244 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:34 crc kubenswrapper[4824]: E1124 13:15:34.009302 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:34 crc kubenswrapper[4824]: E1124 13:15:34.009632 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:34 crc kubenswrapper[4824]: E1124 13:15:34.009757 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.010106 4824 scope.go:117] "RemoveContainer" containerID="1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62" Nov 24 13:15:34 crc kubenswrapper[4824]: E1124 13:15:34.010321 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.012473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.012558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.012569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.012589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.012600 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.115912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.115969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.115981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.116002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.116017 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.218786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.218849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.218858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.218875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.218892 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.321728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.321779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.321788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.321857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.321876 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.426391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.426445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.426454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.426468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.426478 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.528507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.528548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.528559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.528576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.528586 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.631296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.631339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.631353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.631371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.631381 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.734186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.734231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.734245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.734265 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.734274 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.836977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.837024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.837034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.837052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.837069 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.939361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.939412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.939423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.939443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:34 crc kubenswrapper[4824]: I1124 13:15:34.939457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:34Z","lastTransitionTime":"2025-11-24T13:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.009428 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:35 crc kubenswrapper[4824]: E1124 13:15:35.009626 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.043642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.043696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.043709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.043727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.043739 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.147167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.147333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.147349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.147371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.147388 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.251573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.251648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.251662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.251683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.251700 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.355685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.355734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.355744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.355762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.355773 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.458701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.459135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.459146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.459164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.459175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.530742 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/0.log" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.530801 4824 generic.go:334] "Generic (PLEG): container finished" podID="2599a7bb-fad5-4c40-901b-189dc6687928" containerID="42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c" exitCode=1 Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.530841 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ssstr" event={"ID":"2599a7bb-fad5-4c40-901b-189dc6687928","Type":"ContainerDied","Data":"42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.531497 4824 scope.go:117] "RemoveContainer" containerID="42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.547967 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.563482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.563533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.563546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.563563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.563595 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.563953 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.576083 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.589420 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.603506 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c5aeab-efe8-4aac-a669-339d07505417\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.619636 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.637137 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.653828 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.666790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.666853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.666863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.666880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.666890 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.668036 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.683482 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.696179 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"2025-11-24T13:14:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab\\\\n2025-11-24T13:14:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab to /host/opt/cni/bin/\\\\n2025-11-24T13:14:50Z [verbose] multus-daemon started\\\\n2025-11-24T13:14:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:15:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.706192 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.765229 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:19Z\\\",\\\"message\\\":\\\"1.EgressIP event handler 8\\\\nI1124 13:15:19.170559 6461 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:19.170570 6461 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:19.170602 6461 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:15:19.170613 6461 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:19.170619 6461 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:15:19.170626 6461 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:19.170637 6461 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:19.170645 6461 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:15:19.170661 6461 factory.go:656] Stopping watch factory\\\\nI1124 13:15:19.170680 6461 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:15:19.170688 6461 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:15:19.170837 6461 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1124 13:15:19.170920 6461 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:15:19.170958 6461 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:19.170980 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:19.171073 6461 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.770124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.770176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.770188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.770207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.770219 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.786703 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.817789 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.830575 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.842299 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.858592 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.872501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.872581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.872600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.872628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.872652 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.878582 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.976255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.976380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.976399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.976424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:35 crc kubenswrapper[4824]: I1124 13:15:35.976442 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:35Z","lastTransitionTime":"2025-11-24T13:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.010014 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.010026 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:36 crc kubenswrapper[4824]: E1124 13:15:36.010204 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.010580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:36 crc kubenswrapper[4824]: E1124 13:15:36.010756 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:36 crc kubenswrapper[4824]: E1124 13:15:36.011019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.079231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.079279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.079293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.079318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.079332 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:36Z","lastTransitionTime":"2025-11-24T13:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.182329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.182385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.182395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.182416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.182427 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:36Z","lastTransitionTime":"2025-11-24T13:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.284754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.284823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.284839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.284857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.284870 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:36Z","lastTransitionTime":"2025-11-24T13:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.388974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.389011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.389019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.389035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.389045 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:36Z","lastTransitionTime":"2025-11-24T13:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.491741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.491789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.491801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.491832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.491842 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:36Z","lastTransitionTime":"2025-11-24T13:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.536347 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/0.log" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.536395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ssstr" event={"ID":"2599a7bb-fad5-4c40-901b-189dc6687928","Type":"ContainerStarted","Data":"91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.555023 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.568725 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"2025-11-24T13:14:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab\\\\n2025-11-24T13:14:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab to /host/opt/cni/bin/\\\\n2025-11-24T13:14:50Z [verbose] multus-daemon started\\\\n2025-11-24T13:14:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:15:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.581704 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.594276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.594330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.594349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.594365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.594376 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:36Z","lastTransitionTime":"2025-11-24T13:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.595371 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.614857 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.639308 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:19Z\\\",\\\"message\\\":\\\"1.EgressIP event handler 8\\\\nI1124 13:15:19.170559 6461 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:19.170570 6461 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:19.170602 6461 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:15:19.170613 6461 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:19.170619 6461 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:15:19.170626 6461 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:19.170637 6461 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:19.170645 6461 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:15:19.170661 6461 factory.go:656] Stopping watch factory\\\\nI1124 13:15:19.170680 6461 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:15:19.170688 6461 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:15:19.170837 6461 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1124 13:15:19.170920 6461 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:15:19.170958 6461 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:19.170980 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:19.171073 6461 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.650555 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.673292 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.687101 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.696226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.696248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.696258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.696272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.696281 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:36Z","lastTransitionTime":"2025-11-24T13:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.700300 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.712685 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.725353 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.739143 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.753125 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.768754 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.785535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.799793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.799876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.799890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.799912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.799925 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:36Z","lastTransitionTime":"2025-11-24T13:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.801510 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c5aeab-efe8-4aac-a669-339d07505417\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.817498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.833077 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:36Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.902388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.902442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.902451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.902468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:36 crc kubenswrapper[4824]: I1124 13:15:36.902480 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:36Z","lastTransitionTime":"2025-11-24T13:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.005959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.006014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.006025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.006045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.006060 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.009305 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:37 crc kubenswrapper[4824]: E1124 13:15:37.009560 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.109473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.109530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.109547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.109571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.109585 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.212591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.212638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.212647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.212665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.212678 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.316025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.316083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.316101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.316125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.316139 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.419499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.419568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.419585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.419625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.419637 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.523274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.523342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.523354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.523375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.523392 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.625667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.625919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.625930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.625947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.625960 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.728357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.728399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.728410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.728424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.728436 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.832147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.832207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.832218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.832237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.832248 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.934784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.934907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.934927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.934948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:37 crc kubenswrapper[4824]: I1124 13:15:37.934963 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:37Z","lastTransitionTime":"2025-11-24T13:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.009928 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.009931 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:38 crc kubenswrapper[4824]: E1124 13:15:38.010085 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.009952 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:38 crc kubenswrapper[4824]: E1124 13:15:38.010182 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:38 crc kubenswrapper[4824]: E1124 13:15:38.010449 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.037301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.037369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.037389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.037421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.037445 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.140870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.140923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.140938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.140961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.140974 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.244558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.244641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.244668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.244698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.244719 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.332393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.332442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.332469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.332496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.332510 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: E1124 13:15:38.350712 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.357708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.357789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.357860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.357895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.357922 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: E1124 13:15:38.373497 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.378100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.378150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.378161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.378178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.378189 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: E1124 13:15:38.391507 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.395937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.395970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.395981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.395999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.396012 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: E1124 13:15:38.421697 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.427353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.427392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.427404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.427420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.427433 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: E1124 13:15:38.444802 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:38 crc kubenswrapper[4824]: E1124 13:15:38.444988 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.446913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.446992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.447005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.447026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.447039 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.548987 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.549035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.549044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.549060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.549072 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.652256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.652293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.652303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.652320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.652331 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.755071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.755183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.755203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.755233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.755251 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.858475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.858511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.858521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.858537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.858548 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.962037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.962086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.962099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.962118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:38 crc kubenswrapper[4824]: I1124 13:15:38.962133 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:38Z","lastTransitionTime":"2025-11-24T13:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.009979 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:39 crc kubenswrapper[4824]: E1124 13:15:39.010170 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.024854 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.037661 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.053103 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c5aeab-efe8-4aac-a669-339d07505417\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.065228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.065286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.065301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.065324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.065338 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.070640 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.086236 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.100043 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.113969 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.132314 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.147932 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"2025-11-24T13:14:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab\\\\n2025-11-24T13:14:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab to /host/opt/cni/bin/\\\\n2025-11-24T13:14:50Z [verbose] multus-daemon started\\\\n2025-11-24T13:14:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:15:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.162701 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.167891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.167953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.167970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.167992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.168005 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.187668 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:19Z\\\",\\\"message\\\":\\\"1.EgressIP event handler 8\\\\nI1124 13:15:19.170559 6461 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:19.170570 6461 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:19.170602 6461 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:15:19.170613 6461 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:19.170619 6461 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:15:19.170626 6461 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:19.170637 6461 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:19.170645 6461 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:15:19.170661 6461 factory.go:656] Stopping watch factory\\\\nI1124 13:15:19.170680 6461 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:15:19.170688 6461 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:15:19.170837 6461 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1124 13:15:19.170920 6461 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:15:19.170958 6461 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:19.170980 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:19.171073 6461 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.201334 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.229472 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.248302 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.263583 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.270892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.270945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.270954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.270968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.270977 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.278903 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.299612 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.314330 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.327388 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.375265 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.375324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.375338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.375358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.375374 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.478329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.478377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.478386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.478402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.478414 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.582001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.582062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.582075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.582098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.582110 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.685751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.685801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.685836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.685855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.685866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.789010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.789057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.789068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.789085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.789094 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.892269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.892321 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.892339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.892358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.892371 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.994703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.994750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.994767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.994785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:39 crc kubenswrapper[4824]: I1124 13:15:39.994799 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:39Z","lastTransitionTime":"2025-11-24T13:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.009150 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:40 crc kubenswrapper[4824]: E1124 13:15:40.009289 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.009479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:40 crc kubenswrapper[4824]: E1124 13:15:40.009540 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.009655 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:40 crc kubenswrapper[4824]: E1124 13:15:40.009707 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.097640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.097674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.097682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.097697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.097707 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:40Z","lastTransitionTime":"2025-11-24T13:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.200673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.200751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.200774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.200803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.201236 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:40Z","lastTransitionTime":"2025-11-24T13:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.304424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.304483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.304493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.304514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.304524 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:40Z","lastTransitionTime":"2025-11-24T13:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.406960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.407015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.407032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.407055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.407070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:40Z","lastTransitionTime":"2025-11-24T13:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.510312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.510355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.510366 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.510381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.510390 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:40Z","lastTransitionTime":"2025-11-24T13:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.613613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.613658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.613667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.613688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.613699 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:40Z","lastTransitionTime":"2025-11-24T13:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.716009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.716054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.716063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.716079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.716088 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:40Z","lastTransitionTime":"2025-11-24T13:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.819150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.819234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.819246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.819267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.819279 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:40Z","lastTransitionTime":"2025-11-24T13:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.921609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.921679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.921695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.921718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:40 crc kubenswrapper[4824]: I1124 13:15:40.921732 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:40Z","lastTransitionTime":"2025-11-24T13:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.010121 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:41 crc kubenswrapper[4824]: E1124 13:15:41.010324 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.024764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.024824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.024835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.024854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.024866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.127715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.127757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.127770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.127788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.127831 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.230503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.230550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.230562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.230583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.230594 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.333966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.334025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.334035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.334055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.334066 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.436666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.436730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.436742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.436761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.436773 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.539704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.539742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.539758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.539772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.539784 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.642567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.642613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.642624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.642641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.642652 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.747122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.747161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.747172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.747186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.747194 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.850127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.850173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.850183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.850201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.850210 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.952889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.953270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.953359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.953449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:41 crc kubenswrapper[4824]: I1124 13:15:41.953545 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:41Z","lastTransitionTime":"2025-11-24T13:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.009763 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.009871 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:42 crc kubenswrapper[4824]: E1124 13:15:42.010456 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:42 crc kubenswrapper[4824]: E1124 13:15:42.010301 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.009895 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:42 crc kubenswrapper[4824]: E1124 13:15:42.010744 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.057113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.057147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.057159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.057177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.057187 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.160237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.160319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.160341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.160372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.160394 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.265001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.265062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.265073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.265094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.265107 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.368248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.368311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.368327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.368346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.368359 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.471449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.471525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.471543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.471568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.471587 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.573972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.574010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.574021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.574045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.574057 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.678102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.678176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.678200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.678229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.678251 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.780439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.780506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.780519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.780539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.780551 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.883302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.883343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.883353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.883372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.883383 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.986536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.986577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.986588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.986603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:42 crc kubenswrapper[4824]: I1124 13:15:42.986612 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:42Z","lastTransitionTime":"2025-11-24T13:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.010174 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.010353 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.089662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.089710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.089719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.089734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.089744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:43Z","lastTransitionTime":"2025-11-24T13:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.193201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.193288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.193308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.193337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.193357 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:43Z","lastTransitionTime":"2025-11-24T13:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.295962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.296013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.296025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.296044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.296055 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:43Z","lastTransitionTime":"2025-11-24T13:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.399378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.399434 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.399446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.399464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.399473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:43Z","lastTransitionTime":"2025-11-24T13:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.503140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.503194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.503205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.503227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.503241 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:43Z","lastTransitionTime":"2025-11-24T13:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.606440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.606511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.606525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.606560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.606580 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:43Z","lastTransitionTime":"2025-11-24T13:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.709452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.709507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.709530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.709553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.709568 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:43Z","lastTransitionTime":"2025-11-24T13:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.812555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.812649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.812669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.812700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.812731 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:43Z","lastTransitionTime":"2025-11-24T13:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.843285 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.843458 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:47.84342617 +0000 UTC m=+149.482965490 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.843549 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.843616 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.843747 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.843747 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.843765 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.843771 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.843778 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.843783 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.843851 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:16:47.843843618 +0000 UTC m=+149.483382918 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.843864 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:16:47.843859608 +0000 UTC m=+149.483398918 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.915567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.915617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.915627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.915644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.915657 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:43Z","lastTransitionTime":"2025-11-24T13:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.946279 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:43 crc kubenswrapper[4824]: I1124 13:15:43.946481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.946680 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.946783 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:16:47.946754409 +0000 UTC m=+149.586293749 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.947476 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:15:43 crc kubenswrapper[4824]: E1124 13:15:43.947650 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:16:47.947602095 +0000 UTC m=+149.587141425 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.009372 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:44 crc kubenswrapper[4824]: E1124 13:15:44.009556 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.009784 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:44 crc kubenswrapper[4824]: E1124 13:15:44.009941 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.010234 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:44 crc kubenswrapper[4824]: E1124 13:15:44.010513 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.018755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.018791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.018799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.018831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.018847 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.120952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.121008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.121021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.121040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.121055 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.224431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.224463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.224471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.224486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.224497 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.327089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.327143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.327158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.327181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.327193 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.429754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.429798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.429846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.429862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.429874 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.533020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.533082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.533095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.533118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.533131 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.636343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.636398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.636417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.636444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.636463 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.739918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.739964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.739977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.739996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.740008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.843495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.843931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.844108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.844309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.844484 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.947917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.947971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.947985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.948003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:44 crc kubenswrapper[4824]: I1124 13:15:44.948014 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:44Z","lastTransitionTime":"2025-11-24T13:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.010019 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:45 crc kubenswrapper[4824]: E1124 13:15:45.010209 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.051049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.051138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.051153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.051177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.051194 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.154294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.154380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.154398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.154479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.154498 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.258086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.258154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.258168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.258189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.258205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.361736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.361781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.361794 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.361849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.361862 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.466447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.466493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.466503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.466520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.466531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.571575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.571695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.571708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.571735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.571751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.673981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.674038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.674050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.674069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.674080 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.778060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.778121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.778145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.778176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.778230 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.881664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.881721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.881732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.881753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.881769 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.985932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.986144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.986296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.986399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:45 crc kubenswrapper[4824]: I1124 13:15:45.986437 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:45Z","lastTransitionTime":"2025-11-24T13:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.010122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.010490 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:46 crc kubenswrapper[4824]: E1124 13:15:46.010711 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.010191 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:46 crc kubenswrapper[4824]: E1124 13:15:46.011535 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:46 crc kubenswrapper[4824]: E1124 13:15:46.011659 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.012034 4824 scope.go:117] "RemoveContainer" containerID="1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.091285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.091338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.091355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.091376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.091394 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:46Z","lastTransitionTime":"2025-11-24T13:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.194549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.194625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.194644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.194671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.194692 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:46Z","lastTransitionTime":"2025-11-24T13:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.299641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.299694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.299707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.299726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.299738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:46Z","lastTransitionTime":"2025-11-24T13:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.402590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.402969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.402979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.402994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.403004 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:46Z","lastTransitionTime":"2025-11-24T13:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.506069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.506116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.506127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.506146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.506156 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:46Z","lastTransitionTime":"2025-11-24T13:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.577031 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/2.log" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.581207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.582162 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.600488 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.609000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.609073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.609086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.609103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.609118 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:46Z","lastTransitionTime":"2025-11-24T13:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.617316 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.638312 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:19Z\\\",\\\"message\\\":\\\"1.EgressIP event handler 8\\\\nI1124 13:15:19.170559 6461 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:19.170570 6461 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:19.170602 6461 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:15:19.170613 6461 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:19.170619 6461 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:15:19.170626 6461 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:19.170637 6461 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:19.170645 6461 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:15:19.170661 6461 factory.go:656] Stopping watch factory\\\\nI1124 13:15:19.170680 6461 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:15:19.170688 6461 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:15:19.170837 6461 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1124 13:15:19.170920 6461 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:15:19.170958 6461 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:19.170980 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:19.171073 6461 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.652470 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.679383 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.697569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.711565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.711599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.711608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.711622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.711634 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:46Z","lastTransitionTime":"2025-11-24T13:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.715056 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.737374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.757750 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.781345 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.793796 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.803560 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.814679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.814716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.814726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.814740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.814750 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:46Z","lastTransitionTime":"2025-11-24T13:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.814962 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.826500 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c5aeab-efe8-4aac-a669-339d07505417\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.839603 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.865161 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.880827 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.896361 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"2025-11-24T13:14:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab\\\\n2025-11-24T13:14:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab to /host/opt/cni/bin/\\\\n2025-11-24T13:14:50Z [verbose] multus-daemon started\\\\n2025-11-24T13:14:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:15:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.910248 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.916908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.916940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.916951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.916971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:46 crc kubenswrapper[4824]: I1124 13:15:46.916981 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:46Z","lastTransitionTime":"2025-11-24T13:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.010186 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:47 crc kubenswrapper[4824]: E1124 13:15:47.010353 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.019376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.019454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.019463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.019480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.019490 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.122744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.122791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.122824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.122843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.122856 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.226507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.226558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.226570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.226592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.226606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.329069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.329127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.329137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.329158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.329174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.432332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.432415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.432432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.432454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.432467 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.538628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.538686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.538703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.538726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.538741 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.586857 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/3.log" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.588098 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/2.log" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.591483 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42" exitCode=1 Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.591553 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.591622 4824 scope.go:117] "RemoveContainer" containerID="1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.593274 4824 scope.go:117] "RemoveContainer" containerID="377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42" Nov 24 13:15:47 crc kubenswrapper[4824]: E1124 13:15:47.593642 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.616539 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.639138 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b30957a75094ffea4963404269028ed1569cd519074873ff0ef3076c3442f62\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:19Z\\\",\\\"message\\\":\\\"1.EgressIP event handler 8\\\\nI1124 13:15:19.170559 6461 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:15:19.170570 6461 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:15:19.170602 6461 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:15:19.170613 6461 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:19.170619 6461 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:15:19.170626 6461 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:15:19.170637 6461 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:19.170645 6461 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:15:19.170661 6461 factory.go:656] Stopping watch factory\\\\nI1124 13:15:19.170680 6461 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:15:19.170688 6461 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:15:19.170837 6461 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1124 13:15:19.170920 6461 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:15:19.170958 6461 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:19.170980 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:19.171073 6461 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:47Z\\\",\\\"message\\\":\\\"asedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945740 6806 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945860 6806 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:46.945997 6806 factory.go:656] Stopping watch factory\\\\nI1124 13:15:46.946002 6806 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945893 6806 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:15:46.946608 6806 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:46.963175 6806 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1124 13:15:46.963194 6806 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1124 13:15:46.963249 6806 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:46.963271 6806 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:46.963391 6806 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.641686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.641723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.641734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.641751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.641763 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.652895 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.682879 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.695477 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.707232 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.719544 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.734464 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.744119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.744169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.744182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.744198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.744212 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.752968 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.765190 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.775902 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.790778 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.802484 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c5aeab-efe8-4aac-a669-339d07505417\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.816338 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.828908 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.841048 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.847426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.847473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.847484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.847499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.847510 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.858465 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.871099 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"2025-11-24T13:14:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab\\\\n2025-11-24T13:14:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab to /host/opt/cni/bin/\\\\n2025-11-24T13:14:50Z [verbose] multus-daemon started\\\\n2025-11-24T13:14:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:15:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.883519 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.950339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.950375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.950384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.950401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:47 crc kubenswrapper[4824]: I1124 13:15:47.950410 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:47Z","lastTransitionTime":"2025-11-24T13:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.009969 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.010030 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.010071 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.010128 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.010246 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.010381 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.053848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.053911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.053925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.053949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.053964 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.156641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.156689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.156698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.156714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.156723 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.259092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.259156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.259186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.259205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.259217 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.362598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.362657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.362675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.362697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.362712 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.465932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.466202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.466225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.466251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.466270 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.569723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.569778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.569788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.569822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.569839 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.596323 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/3.log" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.600276 4824 scope.go:117] "RemoveContainer" containerID="377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.600503 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.611234 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.624440 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.637923 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"2025-11-24T13:14:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab\\\\n2025-11-24T13:14:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab to /host/opt/cni/bin/\\\\n2025-11-24T13:14:50Z [verbose] multus-daemon started\\\\n2025-11-24T13:14:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:15:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.650259 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.662541 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.675573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.675623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.675641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.675662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.675678 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.691551 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.714603 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:47Z\\\",\\\"message\\\":\\\"asedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945740 6806 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945860 6806 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:46.945997 6806 factory.go:656] Stopping watch factory\\\\nI1124 13:15:46.946002 6806 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945893 6806 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:15:46.946608 6806 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:46.963175 6806 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1124 13:15:46.963194 6806 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1124 13:15:46.963249 6806 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:46.963271 6806 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:46.963391 6806 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.722246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.722343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.722362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.722384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.722399 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.732601 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.737060 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.741171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.741218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.741231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.741251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.741286 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.755913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.757951 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.763868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.763901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.763912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.763930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.763945 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.773047 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.777881 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.781634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.781670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.781683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.781701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.781712 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.790681 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.794485 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.801225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.801272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.801285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.801302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.801314 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.806776 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.814473 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: E1124 13:15:48.814697 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.816743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.816794 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.816826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.816849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.816866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.822406 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.838852 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.850311 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.861683 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.873614 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.883379 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c5aeab-efe8-4aac-a669-339d07505417\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.894794 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.920254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.920296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.920308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.920325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:48 crc kubenswrapper[4824]: I1124 13:15:48.920337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:48Z","lastTransitionTime":"2025-11-24T13:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.009728 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:49 crc kubenswrapper[4824]: E1124 13:15:49.009934 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.022771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.022823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.022835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.022852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.022865 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:49Z","lastTransitionTime":"2025-11-24T13:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.024446 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.035689 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.044790 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.058294 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.069669 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c5aeab-efe8-4aac-a669-339d07505417\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.081309 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.095424 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.107052 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.116770 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.128689 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.143307 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"2025-11-24T13:14:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab\\\\n2025-11-24T13:14:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab to /host/opt/cni/bin/\\\\n2025-11-24T13:14:50Z [verbose] multus-daemon started\\\\n2025-11-24T13:14:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:15:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.157749 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.175268 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:47Z\\\",\\\"message\\\":\\\"asedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945740 6806 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945860 6806 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:46.945997 6806 factory.go:656] Stopping watch factory\\\\nI1124 13:15:46.946002 6806 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945893 6806 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:15:46.946608 6806 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:46.963175 6806 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1124 13:15:46.963194 6806 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1124 13:15:46.963249 6806 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:46.963271 6806 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:46.963391 6806 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.187578 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.991788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.991850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.991866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.992039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:49 crc kubenswrapper[4824]: I1124 13:15:49.992058 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:49Z","lastTransitionTime":"2025-11-24T13:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.011210 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:50 crc kubenswrapper[4824]: E1124 13:15:50.011369 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.011577 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:50 crc kubenswrapper[4824]: E1124 13:15:50.011634 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.011569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:50 crc kubenswrapper[4824]: E1124 13:15:50.011705 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.032380 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.047995 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.062554 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.077800 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.093803 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.094863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.094899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.094908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.094923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.094932 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:50Z","lastTransitionTime":"2025-11-24T13:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.198047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.198087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.198096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.198114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.198124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:50Z","lastTransitionTime":"2025-11-24T13:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.300496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.300546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.300559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.300582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.300595 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:50Z","lastTransitionTime":"2025-11-24T13:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.403333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.403373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.403381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.403396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.403407 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:50Z","lastTransitionTime":"2025-11-24T13:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.506006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.506046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.506056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.506073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.506085 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:50Z","lastTransitionTime":"2025-11-24T13:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.608864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.608944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.608958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.608985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.609002 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:50Z","lastTransitionTime":"2025-11-24T13:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.712551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.712600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.712610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.712631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.712644 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:50Z","lastTransitionTime":"2025-11-24T13:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.816260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.816331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.816342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.816364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.816376 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:50Z","lastTransitionTime":"2025-11-24T13:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.918854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.918900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.918914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.918935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:50 crc kubenswrapper[4824]: I1124 13:15:50.918950 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:50Z","lastTransitionTime":"2025-11-24T13:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.009517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:51 crc kubenswrapper[4824]: E1124 13:15:51.009718 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.020995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.021496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.021673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.021879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.022032 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.125496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.125550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.125567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.125592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.125606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.228503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.228580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.228604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.228641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.228664 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.332553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.332609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.332619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.332640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.332650 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.435603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.435657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.435666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.435681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.435718 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.539016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.539070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.539079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.539096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.539107 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.641185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.641233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.641242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.641257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.641268 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.743754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.743884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.743928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.743966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.743988 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.847467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.847534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.847552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.847577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.847595 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.951302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.951376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.951400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.951433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:51 crc kubenswrapper[4824]: I1124 13:15:51.951456 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:51Z","lastTransitionTime":"2025-11-24T13:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.009280 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.009329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.009476 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:52 crc kubenswrapper[4824]: E1124 13:15:52.009490 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:52 crc kubenswrapper[4824]: E1124 13:15:52.009700 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:52 crc kubenswrapper[4824]: E1124 13:15:52.009779 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.055107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.055157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.055169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.055190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.055205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.157859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.157951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.157963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.157981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.157993 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.260602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.261061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.261072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.261088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.261099 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.363908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.364317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.364649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.365107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.365334 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.469238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.469634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.469771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.469946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.470062 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.572839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.572886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.572896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.572912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.572924 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.675342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.675743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.675913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.676044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.676304 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.779356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.779422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.779435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.779455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.779474 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.881481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.881515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.881524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.881540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.881551 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.984039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.984067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.984075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.984088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:52 crc kubenswrapper[4824]: I1124 13:15:52.984097 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:52Z","lastTransitionTime":"2025-11-24T13:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.009057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:53 crc kubenswrapper[4824]: E1124 13:15:53.009189 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.087174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.087214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.087224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.087238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.087247 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:53Z","lastTransitionTime":"2025-11-24T13:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.190390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.190449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.190465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.190489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.190505 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:53Z","lastTransitionTime":"2025-11-24T13:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.293347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.293383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.293391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.293405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.293415 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:53Z","lastTransitionTime":"2025-11-24T13:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.396392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.396436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.396444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.396458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.396467 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:53Z","lastTransitionTime":"2025-11-24T13:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.499973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.500050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.500075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.500106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.500128 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:53Z","lastTransitionTime":"2025-11-24T13:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.603373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.603419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.603430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.603447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.603457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:53Z","lastTransitionTime":"2025-11-24T13:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.706175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.706221 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.706235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.706252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.706263 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:53Z","lastTransitionTime":"2025-11-24T13:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.809602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.809643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.809651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.809666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.809675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:53Z","lastTransitionTime":"2025-11-24T13:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.912872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.912962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.912993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.913024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:53 crc kubenswrapper[4824]: I1124 13:15:53.913048 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:53Z","lastTransitionTime":"2025-11-24T13:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.010051 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.010170 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:54 crc kubenswrapper[4824]: E1124 13:15:54.010227 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.010278 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:54 crc kubenswrapper[4824]: E1124 13:15:54.010384 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:54 crc kubenswrapper[4824]: E1124 13:15:54.010462 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.016913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.016942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.016954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.016971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.016983 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.120097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.120140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.120170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.120190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.120205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.223920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.223959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.223970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.223988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.223999 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.327434 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.327541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.327563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.327626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.327649 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.431030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.431088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.431100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.431120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.431132 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.534095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.534141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.534156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.534178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.534192 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.636743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.636799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.636833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.636847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.636856 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.739070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.739130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.739152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.739185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.739211 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.841691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.841744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.841755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.841774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.841785 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.944185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.944235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.944244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.944260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:54 crc kubenswrapper[4824]: I1124 13:15:54.944271 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:54Z","lastTransitionTime":"2025-11-24T13:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.009606 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:55 crc kubenswrapper[4824]: E1124 13:15:55.009838 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.046541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.046573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.046580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.046593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.046604 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.149684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.149715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.149723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.149736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.149745 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.252871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.252933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.252941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.252956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.252966 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.355396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.355434 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.355446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.355460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.355469 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.458868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.458924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.458941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.458968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.458986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.561541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.561590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.561606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.561627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.561645 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.664242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.664275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.664309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.664324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.664334 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.768922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.768991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.769006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.769027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.769043 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.871595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.871636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.871647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.871664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.871674 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.977079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.977116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.977130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.977149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:55 crc kubenswrapper[4824]: I1124 13:15:55.977162 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:55Z","lastTransitionTime":"2025-11-24T13:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.010083 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.010144 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:56 crc kubenswrapper[4824]: E1124 13:15:56.010229 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:56 crc kubenswrapper[4824]: E1124 13:15:56.010326 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.010427 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:56 crc kubenswrapper[4824]: E1124 13:15:56.010507 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.079108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.079159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.079170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.079185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.079194 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:56Z","lastTransitionTime":"2025-11-24T13:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.181609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.181656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.181667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.181683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.181694 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:56Z","lastTransitionTime":"2025-11-24T13:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.287654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.287720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.287737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.287762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.287780 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:56Z","lastTransitionTime":"2025-11-24T13:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.394799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.394868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.394880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.394898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.394910 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:56Z","lastTransitionTime":"2025-11-24T13:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.497498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.497549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.497572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.497600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.497616 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:56Z","lastTransitionTime":"2025-11-24T13:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.600578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.600622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.600635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.600686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.600725 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:56Z","lastTransitionTime":"2025-11-24T13:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.704214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.704260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.704278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.704299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.704311 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:56Z","lastTransitionTime":"2025-11-24T13:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.807258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.807348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.807364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.807385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.807400 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:56Z","lastTransitionTime":"2025-11-24T13:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.910159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.910373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.910415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.910528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:56 crc kubenswrapper[4824]: I1124 13:15:56.910557 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:56Z","lastTransitionTime":"2025-11-24T13:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.009537 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:57 crc kubenswrapper[4824]: E1124 13:15:57.009722 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.013442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.013515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.013530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.013548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.013563 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.117404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.118223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.118280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.118310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.118328 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.221644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.221735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.221779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.221850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.221869 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.325311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.325374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.325386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.325424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.325435 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.429594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.429703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.429803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.429881 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.429898 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.532575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.532622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.532631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.532644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.532653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.635910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.635990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.636022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.636070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.636095 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.739224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.739275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.739287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.739306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.739319 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.843035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.843101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.843124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.843154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.843177 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.946090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.946141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.946152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.946168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:57 crc kubenswrapper[4824]: I1124 13:15:57.946179 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:57Z","lastTransitionTime":"2025-11-24T13:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.009794 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.009873 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:15:58 crc kubenswrapper[4824]: E1124 13:15:58.009975 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:15:58 crc kubenswrapper[4824]: E1124 13:15:58.010081 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.010329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:15:58 crc kubenswrapper[4824]: E1124 13:15:58.010410 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.048478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.048528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.048560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.048581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.048594 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.151646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.151728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.151932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.151994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.152007 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.255416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.255465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.255480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.255501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.255516 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.357595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.357653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.357665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.357683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.357696 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.460781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.460874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.460886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.460904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.460917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.563619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.563673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.563689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.563711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.563726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.666965 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.667024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.667038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.667054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.667065 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.769313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.769375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.769396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.769420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.769439 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.872375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.872433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.872452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.872482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.872499 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.975735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.975776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.975787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.975843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:58 crc kubenswrapper[4824]: I1124 13:15:58.975857 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:58Z","lastTransitionTime":"2025-11-24T13:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.010376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:15:59 crc kubenswrapper[4824]: E1124 13:15:59.010663 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.013065 4824 scope.go:117] "RemoveContainer" containerID="377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42" Nov 24 13:15:59 crc kubenswrapper[4824]: E1124 13:15:59.013662 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.033709 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38c30c02-6e81-42bd-bae3-34cd1c57567b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510ba1a50184b73b2a197220f50518ffc22f5dc766db554ecbcb4e258843871c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e348fc72fb69c02031ae24b57fbc74217b0841627f1c1717660530288f3f9b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c49e4d361d35663cfc38d0ed63dc8c38a0e1c6c5efd3afd7c1bd542f34e23a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254600802acb7a805e29f8ef660ebe4b7e3461b3707043e54f423dfb2e98135d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d19290da492d2520c8fa541137ee87869833c42048335084f9a9f497a392e97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 13:14:39.909457 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 13:14:39.909764 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 13:14:39.910725 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1310313645/tls.crt::/tmp/serving-cert-1310313645/tls.key\\\\\\\"\\\\nI1124 13:14:40.254992 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 13:14:40.257633 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 13:14:40.257653 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 13:14:40.257676 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 13:14:40.257683 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 13:14:40.263646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 13:14:40.263673 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263679 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 13:14:40.263685 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 13:14:40.263689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 13:14:40.263692 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 13:14:40.263695 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 13:14:40.263780 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 13:14:40.264857 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c3b2b69e934a9862e36dc080fac51c76357477bcc17b5dd5b567f3ec6a2985f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efd2927cca88013aa42aa79ba923e62fbcb5cda6d1980ffd07ff2989b9185acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.058996 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ssstr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2599a7bb-fad5-4c40-901b-189dc6687928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:35Z\\\",\\\"message\\\":\\\"2025-11-24T13:14:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab\\\\n2025-11-24T13:14:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_11e43923-8e74-43eb-bfd2-41068dfb18ab to /host/opt/cni/bin/\\\\n2025-11-24T13:14:50Z [verbose] multus-daemon started\\\\n2025-11-24T13:14:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:15:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwlhd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ssstr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.079249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.079315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.079337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.079367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.079389 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.079940 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b90f50d7-5574-472e-9532-91b491a0cc56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5645edd66be57ee40a5026756915c96206aeaa5bebd5e34628f9c07d2bb832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfdfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-74f4n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.096789 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sc887" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e81fafe-6567-47f5-bb87-78fffb25fa9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42w7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:15:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sc887\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.120684 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bc9887b-30b9-4cea-8ab9-75b7fe2989b3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d63d22e65414317f6a85cc636b35471c05d64d14222c1e123455221c26fbf60f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b1ede25a2ba0182b6af84ab47f25851b79814d076695332a447386a421d89c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a599b41adefc22ef73163fea96316f934c93a399d8fa6e1929c6924e4032a42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a6f390bbe3f968158a3191c6eeebfe1df0bff0275d5d290f38362cfee87b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0941d76cd354181f9f2a25b08c428a0a4e21b635d138b267f5280658ef6c5e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589f9181349f4ad4c65f3d9b32813912a5dd065b6d114bf0902011ed238fd8f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36fdadc76c8d1d3d1931124e05315cc77b9cd962df0e3e27f5f1edfb85479a7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57a78d86dffaade71cd8c88e6d86abb3af65e36c913ffa747dbd1683038d77a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.122323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.122397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.122465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.122508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.122537 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: E1124 13:15:59.143998 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.148241 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.149802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.149910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.149934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.149972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.149994 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.166130 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: E1124 13:15:59.168440 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.174413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.174470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.174488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.174509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.174523 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.185742 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cc1d2a222986609bdc87d3085a58ec3429c4a7fafe5d76426daf8e771faa63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a82c9f2ce997f2a55e99bc2d29d5e235aaea4304765d4a8b7465087c209f53a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: E1124 13:15:59.188750 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.196118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.196340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.196432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.196529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.196616 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.202319 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc1fdc8e-08cd-4deb-b00c-8b84f656a6d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f46dd7983c12c5a5f3f628efa14fd0290e4462783b34000963fe357166409a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e01428c832d2ee73ef31f28830ecd62a3f14fafe5e59fc6c7dd2330bc5fae21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84dad539069286b626c93734f8206ec873005b95fbdd262b403dfc4e864ba4d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b13b04fa8f9a9879153f1d8cc1fe68df767b38fd40178a2bdb06f67c43f56002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32bb05e810c0cfb42394af21dc5c3555745d7a3d97537caa1d0d4b961a7157b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99e280cdb5e01ea111f6ec8a5b6b63623d710a74a17d6202fd0f258566dab80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc524ace0a9402c6fb1b11f27a3fb1d0b7873373d955a8a7482abe266e1cf733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzlnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wxxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: E1124 13:15:59.211627 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.215340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.215365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.215374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.215386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.215396 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.225035 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f43932f-b763-4fea-9ed9-737078e6548d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:15:47Z\\\",\\\"message\\\":\\\"asedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945740 6806 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945860 6806 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:15:46.945997 6806 factory.go:656] Stopping watch factory\\\\nI1124 13:15:46.946002 6806 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:15:46.945893 6806 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:15:46.946608 6806 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 13:15:46.963175 6806 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1124 13:15:46.963194 6806 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1124 13:15:46.963249 6806 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:15:46.963271 6806 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:15:46.963391 6806 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:15:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8vgzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nk5qz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: E1124 13:15:59.229054 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50fc599-297b-4acb-83a1-35b8955262d5\\\",\\\"systemUUID\\\":\\\"8f1e113f-2cfe-4be7-935e-94a6d127c173\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: E1124 13:15:59.229428 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.230929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.230948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.230957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.230970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.230978 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.236686 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32fa37c8-fb2b-4475-b35f-12a7766b88da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb5da41e55afe49452d94e79b45a463bbd83c7e8e51450d990a9eea6186bb0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93788d7bb8cb1de9d9b4e66a4745351670834b5d3c6017c6629a36ee561cf31c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e085f26fddc881dee6159daead1bb703c340ce2471d4cd861d4836d34c1a7027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cbe70786982ac3e0a28e3febd744c12d83f9dc30d5c612fe44a1565daf00a06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.246768 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6a16de2de52aa9b92a0da54f1e82319172a7c709be48e1c46a94f6fc5de073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.258334 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff0fb887-4240-4e3d-9844-cdfa8ce2d974\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a52e0d3f6f5eef40e633e62e9412ecc58e870e6056034c03f035a24f5462eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf9652b604380b572dbaa820f992c1e98eeaa69fb1d53ff7cd57934b655e5d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-249dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ppkzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.268449 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c5aeab-efe8-4aac-a669-339d07505417\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e35d76a86742e86c25aa050d1cdaeacb4340ec8250d7a0de8213a10edf82a072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56dba6d45bba167d1667f62506a3ebd6c7b615b836e20a4dc38037c78bcb2711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.285976 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1302017-0684-4ecd-9992-71561d8b136a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac7c93fbabd1b5840be11c369acd37f73b729c46c6f8c8b050d3c34a38bea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2be9e3ca001a085c29bd11fa83ff820b7c8a616b7b3edfc2dffed25b486de6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8c9a3f4faf6062e552b7fdf3bed1a1e17fc70b34f044a80f882b0c2c5c588c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b438e2c0358a9c65511f5d5f54d4b75abd24eacac297412f5be065b8abf4a97c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:14:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:14:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.299358 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1269f6f10751d425a9c1f4819d0de02e8e739001641c9535eddb760eb96ecf76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.313166 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.330003 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xtlj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f9693d-cfd5-4c65-af6a-45d54eb2f5c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a92a2fd3d51253555ca226cb241789f848cda6bc5cbaeb7a6d93a12834d4999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mpbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xtlj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.334411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.334460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.334472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.334492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.334506 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.345631 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lfmjc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e986e9a8-5aca-46af-a68c-b871a678fb37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17d12c755e593ec53205670f20fc828c5c77a3b06416bfdba761db42a84dd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:14:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t45j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:14:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lfmjc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.437722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.437761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.437772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.437788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.437801 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.540661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.540707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.540719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.540734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.540746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.642786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.642842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.642850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.642863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.642874 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.745564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.745642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.745651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.745667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.745677 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.848663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.848749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.848889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.848926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.848954 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.952126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.952194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.952222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.952248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:15:59 crc kubenswrapper[4824]: I1124 13:15:59.952267 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:15:59Z","lastTransitionTime":"2025-11-24T13:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.009240 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.009287 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.009380 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:00 crc kubenswrapper[4824]: E1124 13:16:00.009677 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:00 crc kubenswrapper[4824]: E1124 13:16:00.009436 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:00 crc kubenswrapper[4824]: E1124 13:16:00.009846 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.054828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.055145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.055263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.055401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.055487 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.158206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.158253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.158264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.158278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.158287 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.261083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.261118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.261126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.261141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.261152 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.364228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.364276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.364285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.364301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.364312 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.466216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.466276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.466290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.466316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.466329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.568660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.568712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.568726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.568740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.568750 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.671296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.671346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.671359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.671376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.671386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.774352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.774382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.774390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.774405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.774414 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.876921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.876964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.876976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.876991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.877005 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.979425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.979462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.979473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.979489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:00 crc kubenswrapper[4824]: I1124 13:16:00.979499 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:00Z","lastTransitionTime":"2025-11-24T13:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.009997 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:01 crc kubenswrapper[4824]: E1124 13:16:01.010130 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.082691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.082763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.082778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.082798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.082836 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:01Z","lastTransitionTime":"2025-11-24T13:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.185639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.185669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.185681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.185696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.185704 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:01Z","lastTransitionTime":"2025-11-24T13:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.287983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.288065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.288092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.288156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.288170 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:01Z","lastTransitionTime":"2025-11-24T13:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.390486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.390844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.390961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.391058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.391161 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:01Z","lastTransitionTime":"2025-11-24T13:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.494159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.494513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.494612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.494734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.494924 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:01Z","lastTransitionTime":"2025-11-24T13:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.598674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.599378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.599578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.599926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.600135 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:01Z","lastTransitionTime":"2025-11-24T13:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.702946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.702986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.702998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.703016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.703028 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:01Z","lastTransitionTime":"2025-11-24T13:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.806156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.806504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.806599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.806736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.806902 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:01Z","lastTransitionTime":"2025-11-24T13:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.909597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.909961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.910145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.910258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:01 crc kubenswrapper[4824]: I1124 13:16:01.910345 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:01Z","lastTransitionTime":"2025-11-24T13:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.010170 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.010191 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.010241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:02 crc kubenswrapper[4824]: E1124 13:16:02.010756 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:02 crc kubenswrapper[4824]: E1124 13:16:02.011102 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:02 crc kubenswrapper[4824]: E1124 13:16:02.011183 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.012671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.012721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.012738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.012760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.012784 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.114898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.114976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.115002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.115034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.115057 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.217983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.218027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.218044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.218060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.218070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.320687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.320722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.320729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.320743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.320751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.423803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.423861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.423875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.423890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.423899 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.526265 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.526577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.526679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.526778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.526914 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.629617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.629662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.629672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.629690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.629700 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.732411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.732453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.732464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.732478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.732487 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.835274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.835380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.835748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.836040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.836103 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.938865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.938904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.938913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.938931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:02 crc kubenswrapper[4824]: I1124 13:16:02.938941 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:02Z","lastTransitionTime":"2025-11-24T13:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.008994 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:03 crc kubenswrapper[4824]: E1124 13:16:03.009173 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.041238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.041323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.041347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.041378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.041402 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.144590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.144630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.144639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.144654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.144666 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.247150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.247191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.247206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.247223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.247235 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.349754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.349799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.349823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.349840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.349852 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.452621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.452721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.452746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.452843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.452866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.555569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.555626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.555647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.555683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.555740 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.658567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.658638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.658663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.658688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.658710 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.762182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.762237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.762249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.762266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.762278 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.864777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.864852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.864866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.864883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.864902 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.967563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.967642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.967651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.967670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:03 crc kubenswrapper[4824]: I1124 13:16:03.967680 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:03Z","lastTransitionTime":"2025-11-24T13:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.009698 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.009774 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.009733 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:04 crc kubenswrapper[4824]: E1124 13:16:04.010022 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:04 crc kubenswrapper[4824]: E1124 13:16:04.010118 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:04 crc kubenswrapper[4824]: E1124 13:16:04.010242 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.069150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.069182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.069190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.069205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.069214 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.171309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.171344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.171353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.171366 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.171374 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.274104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.274148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.274157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.274171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.274181 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.376458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.376492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.376502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.376518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.376531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.479632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.479676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.479685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.479704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.479713 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.548006 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:04 crc kubenswrapper[4824]: E1124 13:16:04.548215 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:16:04 crc kubenswrapper[4824]: E1124 13:16:04.548328 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs podName:7e81fafe-6567-47f5-bb87-78fffb25fa9b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:08.548304379 +0000 UTC m=+170.187843679 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs") pod "network-metrics-daemon-sc887" (UID: "7e81fafe-6567-47f5-bb87-78fffb25fa9b") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.582373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.582437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.582447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.582462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.582473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.684866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.684930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.684946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.684968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.684982 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.787262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.787305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.787317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.787333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.787345 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.890953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.891048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.891087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.891129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.891152 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.993831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.994609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.994653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.994679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:04 crc kubenswrapper[4824]: I1124 13:16:04.994697 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:04Z","lastTransitionTime":"2025-11-24T13:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.009726 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:05 crc kubenswrapper[4824]: E1124 13:16:05.009901 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.097322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.097371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.097386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.097408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.097420 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:05Z","lastTransitionTime":"2025-11-24T13:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.200515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.200581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.200600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.200626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.200645 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:05Z","lastTransitionTime":"2025-11-24T13:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.303500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.303588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.303607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.303624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.303634 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:05Z","lastTransitionTime":"2025-11-24T13:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.405939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.406008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.406026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.406054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.406138 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:05Z","lastTransitionTime":"2025-11-24T13:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.508996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.509049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.509108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.509132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.509155 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:05Z","lastTransitionTime":"2025-11-24T13:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.611667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.611739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.611761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.611782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.611793 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:05Z","lastTransitionTime":"2025-11-24T13:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.715513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.715584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.715602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.715657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.715678 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:05Z","lastTransitionTime":"2025-11-24T13:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.818141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.818186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.818197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.818213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.818225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:05Z","lastTransitionTime":"2025-11-24T13:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.921210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.921251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.921261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.921276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:05 crc kubenswrapper[4824]: I1124 13:16:05.921287 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:05Z","lastTransitionTime":"2025-11-24T13:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.009751 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.009790 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.009840 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:06 crc kubenswrapper[4824]: E1124 13:16:06.009965 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:06 crc kubenswrapper[4824]: E1124 13:16:06.010077 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:06 crc kubenswrapper[4824]: E1124 13:16:06.010174 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.023944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.024021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.024050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.024074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.024092 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.126429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.126846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.127053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.127183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.127277 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.230523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.230570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.230580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.230595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.230604 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.334219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.334353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.334376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.334400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.334469 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.438514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.438579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.438605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.438731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.438760 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.541671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.541859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.541876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.541894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.541905 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.644865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.644962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.644989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.645022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.645045 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.748532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.748625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.748650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.748683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.748705 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.851155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.851199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.851211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.851230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.851248 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.954659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.954695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.954705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.954723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:06 crc kubenswrapper[4824]: I1124 13:16:06.954736 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:06Z","lastTransitionTime":"2025-11-24T13:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.009722 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:07 crc kubenswrapper[4824]: E1124 13:16:07.009903 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.057333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.057372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.057381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.057395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.057404 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.160671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.160723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.160736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.160754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.160767 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.264160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.264208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.264217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.264231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.264241 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.367490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.367677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.367699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.367865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.367890 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.471293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.471341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.471354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.471373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.471388 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.574416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.574456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.574466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.574481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.574494 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.676893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.677083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.677110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.677167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.677188 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.780156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.780246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.780264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.780290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.780335 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.885285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.885338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.885352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.885369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.885382 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.987781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.987835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.987850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.987865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:07 crc kubenswrapper[4824]: I1124 13:16:07.987876 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:07Z","lastTransitionTime":"2025-11-24T13:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.009619 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.009648 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.009622 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:08 crc kubenswrapper[4824]: E1124 13:16:08.009761 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:08 crc kubenswrapper[4824]: E1124 13:16:08.009908 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:08 crc kubenswrapper[4824]: E1124 13:16:08.009975 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.089863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.089907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.089922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.089940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.089952 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:08Z","lastTransitionTime":"2025-11-24T13:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.192481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.192522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.192530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.192550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.192569 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:08Z","lastTransitionTime":"2025-11-24T13:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.295012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.295069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.295083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.295099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.295110 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:08Z","lastTransitionTime":"2025-11-24T13:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.397199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.397236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.397247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.397264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.397274 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:08Z","lastTransitionTime":"2025-11-24T13:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.500419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.500509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.500522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.500537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.500567 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:08Z","lastTransitionTime":"2025-11-24T13:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.604077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.604201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.604225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.604256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.604279 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:08Z","lastTransitionTime":"2025-11-24T13:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.707434 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.707490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.707503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.707522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.707536 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:08Z","lastTransitionTime":"2025-11-24T13:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.810574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.810619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.810631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.810652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.810666 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:08Z","lastTransitionTime":"2025-11-24T13:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.913623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.913660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.913670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.913688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:08 crc kubenswrapper[4824]: I1124 13:16:08.913733 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:08Z","lastTransitionTime":"2025-11-24T13:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.011458 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:09 crc kubenswrapper[4824]: E1124 13:16:09.011568 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.016085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.016138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.016150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.016162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.016171 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:09Z","lastTransitionTime":"2025-11-24T13:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.047224 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-xtlj7" podStartSLOduration=84.04720084 podStartE2EDuration="1m24.04720084s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.031936574 +0000 UTC m=+110.671475914" watchObservedRunningTime="2025-11-24 13:16:09.04720084 +0000 UTC m=+110.686740160" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.047551 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-lfmjc" podStartSLOduration=84.047544059 podStartE2EDuration="1m24.047544059s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.046470311 +0000 UTC m=+110.686009621" watchObservedRunningTime="2025-11-24 13:16:09.047544059 +0000 UTC m=+110.687083379" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.076762 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=44.076736896 podStartE2EDuration="44.076736896s" podCreationTimestamp="2025-11-24 13:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.07610539 +0000 UTC m=+110.715644710" watchObservedRunningTime="2025-11-24 13:16:09.076736896 +0000 UTC m=+110.716276226" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.077340 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ppkzl" podStartSLOduration=83.077330471 podStartE2EDuration="1m23.077330471s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.063537202 +0000 UTC m=+110.703076532" watchObservedRunningTime="2025-11-24 13:16:09.077330471 +0000 UTC m=+110.716869791" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.117965 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.117994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.118003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.118016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.118026 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:09Z","lastTransitionTime":"2025-11-24T13:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.118416 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=55.118397378 podStartE2EDuration="55.118397378s" podCreationTimestamp="2025-11-24 13:15:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.098209238 +0000 UTC m=+110.737748588" watchObservedRunningTime="2025-11-24 13:16:09.118397378 +0000 UTC m=+110.757936698" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.169523 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.169502929 podStartE2EDuration="1m28.169502929s" podCreationTimestamp="2025-11-24 13:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.156711046 +0000 UTC m=+110.796250356" watchObservedRunningTime="2025-11-24 13:16:09.169502929 +0000 UTC m=+110.809042239" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.181884 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-ssstr" podStartSLOduration=84.181867902 podStartE2EDuration="1m24.181867902s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.170106244 +0000 UTC m=+110.809645554" watchObservedRunningTime="2025-11-24 13:16:09.181867902 +0000 UTC m=+110.821407212" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.196148 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podStartSLOduration=84.196129802 podStartE2EDuration="1m24.196129802s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.182544219 +0000 UTC m=+110.822083529" watchObservedRunningTime="2025-11-24 13:16:09.196129802 +0000 UTC m=+110.835669112" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.220689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.220725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.220737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.220754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.220766 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:09Z","lastTransitionTime":"2025-11-24T13:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.222698 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9wxxt" podStartSLOduration=84.222683012 podStartE2EDuration="1m24.222683012s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.19684993 +0000 UTC m=+110.836389240" watchObservedRunningTime="2025-11-24 13:16:09.222683012 +0000 UTC m=+110.862222342" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.262578 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=86.26256356 podStartE2EDuration="1m26.26256356s" podCreationTimestamp="2025-11-24 13:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.262396806 +0000 UTC m=+110.901936116" watchObservedRunningTime="2025-11-24 13:16:09.26256356 +0000 UTC m=+110.902102870" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.322769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.322798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.322821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.322835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.322843 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:09Z","lastTransitionTime":"2025-11-24T13:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.373602 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=89.373576363 podStartE2EDuration="1m29.373576363s" podCreationTimestamp="2025-11-24 13:14:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:09.370863515 +0000 UTC m=+111.010402835" watchObservedRunningTime="2025-11-24 13:16:09.373576363 +0000 UTC m=+111.013115673" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.424915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.424954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.424964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.424980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.424990 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:09Z","lastTransitionTime":"2025-11-24T13:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.520087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.520126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.520135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.520149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.520158 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:16:09Z","lastTransitionTime":"2025-11-24T13:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.564254 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r"] Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.564700 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.566801 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.566852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.567102 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.567118 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.704271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/327a3d31-59ae-406f-a639-2151e9ff4053-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.704322 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/327a3d31-59ae-406f-a639-2151e9ff4053-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.704353 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/327a3d31-59ae-406f-a639-2151e9ff4053-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.704496 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/327a3d31-59ae-406f-a639-2151e9ff4053-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.704599 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/327a3d31-59ae-406f-a639-2151e9ff4053-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.805651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/327a3d31-59ae-406f-a639-2151e9ff4053-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.805701 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/327a3d31-59ae-406f-a639-2151e9ff4053-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.805727 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/327a3d31-59ae-406f-a639-2151e9ff4053-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.805774 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/327a3d31-59ae-406f-a639-2151e9ff4053-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.805790 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/327a3d31-59ae-406f-a639-2151e9ff4053-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.805927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/327a3d31-59ae-406f-a639-2151e9ff4053-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.805920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/327a3d31-59ae-406f-a639-2151e9ff4053-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.806889 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/327a3d31-59ae-406f-a639-2151e9ff4053-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.820667 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/327a3d31-59ae-406f-a639-2151e9ff4053-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.823144 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/327a3d31-59ae-406f-a639-2151e9ff4053-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nbq6r\" (UID: \"327a3d31-59ae-406f-a639-2151e9ff4053\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:09 crc kubenswrapper[4824]: I1124 13:16:09.881065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" Nov 24 13:16:10 crc kubenswrapper[4824]: I1124 13:16:10.009116 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:10 crc kubenswrapper[4824]: I1124 13:16:10.009187 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:10 crc kubenswrapper[4824]: I1124 13:16:10.009208 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:10 crc kubenswrapper[4824]: E1124 13:16:10.009552 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:10 crc kubenswrapper[4824]: E1124 13:16:10.009620 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:10 crc kubenswrapper[4824]: E1124 13:16:10.009670 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:10 crc kubenswrapper[4824]: I1124 13:16:10.076421 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" event={"ID":"327a3d31-59ae-406f-a639-2151e9ff4053","Type":"ContainerStarted","Data":"86316b45182a39db0bb3a29c252bc50f7b4f3e3cdcff6c3c42682ecd3d848331"} Nov 24 13:16:10 crc kubenswrapper[4824]: I1124 13:16:10.076468 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" event={"ID":"327a3d31-59ae-406f-a639-2151e9ff4053","Type":"ContainerStarted","Data":"d65f280fdae7e3beed62cce07f956d46b380ff7475f7cdf3a3a1fcde92167df1"} Nov 24 13:16:11 crc kubenswrapper[4824]: I1124 13:16:11.009471 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:11 crc kubenswrapper[4824]: E1124 13:16:11.009600 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:11 crc kubenswrapper[4824]: I1124 13:16:11.094031 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nbq6r" podStartSLOduration=86.094007951 podStartE2EDuration="1m26.094007951s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:11.091537429 +0000 UTC m=+112.731076759" watchObservedRunningTime="2025-11-24 13:16:11.094007951 +0000 UTC m=+112.733547271" Nov 24 13:16:12 crc kubenswrapper[4824]: I1124 13:16:12.010063 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:12 crc kubenswrapper[4824]: I1124 13:16:12.010111 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:12 crc kubenswrapper[4824]: I1124 13:16:12.010063 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:12 crc kubenswrapper[4824]: E1124 13:16:12.010599 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:12 crc kubenswrapper[4824]: E1124 13:16:12.010721 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:12 crc kubenswrapper[4824]: E1124 13:16:12.011221 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:12 crc kubenswrapper[4824]: I1124 13:16:12.011411 4824 scope.go:117] "RemoveContainer" containerID="377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42" Nov 24 13:16:12 crc kubenswrapper[4824]: E1124 13:16:12.011598 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:16:13 crc kubenswrapper[4824]: I1124 13:16:13.009985 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:13 crc kubenswrapper[4824]: E1124 13:16:13.010580 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:14 crc kubenswrapper[4824]: I1124 13:16:14.009025 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:14 crc kubenswrapper[4824]: E1124 13:16:14.009429 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:14 crc kubenswrapper[4824]: I1124 13:16:14.009214 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:14 crc kubenswrapper[4824]: E1124 13:16:14.009670 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:14 crc kubenswrapper[4824]: I1124 13:16:14.009175 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:14 crc kubenswrapper[4824]: E1124 13:16:14.009892 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:15 crc kubenswrapper[4824]: I1124 13:16:15.010413 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:15 crc kubenswrapper[4824]: E1124 13:16:15.010525 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:16 crc kubenswrapper[4824]: I1124 13:16:16.009379 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:16 crc kubenswrapper[4824]: E1124 13:16:16.009554 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:16 crc kubenswrapper[4824]: I1124 13:16:16.009643 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:16 crc kubenswrapper[4824]: E1124 13:16:16.009936 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:16 crc kubenswrapper[4824]: I1124 13:16:16.009942 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:16 crc kubenswrapper[4824]: E1124 13:16:16.010135 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:17 crc kubenswrapper[4824]: I1124 13:16:17.010183 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:17 crc kubenswrapper[4824]: E1124 13:16:17.010334 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:18 crc kubenswrapper[4824]: I1124 13:16:18.009758 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:18 crc kubenswrapper[4824]: I1124 13:16:18.009777 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:18 crc kubenswrapper[4824]: E1124 13:16:18.009928 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:18 crc kubenswrapper[4824]: I1124 13:16:18.010058 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:18 crc kubenswrapper[4824]: E1124 13:16:18.010181 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:18 crc kubenswrapper[4824]: E1124 13:16:18.010211 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:19 crc kubenswrapper[4824]: I1124 13:16:19.010018 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:19 crc kubenswrapper[4824]: E1124 13:16:19.010349 4824 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 13:16:19 crc kubenswrapper[4824]: E1124 13:16:19.014305 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:19 crc kubenswrapper[4824]: E1124 13:16:19.996234 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:16:20 crc kubenswrapper[4824]: I1124 13:16:20.009379 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:20 crc kubenswrapper[4824]: I1124 13:16:20.009998 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:20 crc kubenswrapper[4824]: E1124 13:16:20.010154 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:20 crc kubenswrapper[4824]: I1124 13:16:20.010204 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:20 crc kubenswrapper[4824]: E1124 13:16:20.010554 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:20 crc kubenswrapper[4824]: E1124 13:16:20.012402 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:21 crc kubenswrapper[4824]: I1124 13:16:21.009947 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:21 crc kubenswrapper[4824]: E1124 13:16:21.010148 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:21 crc kubenswrapper[4824]: I1124 13:16:21.113061 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/1.log" Nov 24 13:16:21 crc kubenswrapper[4824]: I1124 13:16:21.113684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/0.log" Nov 24 13:16:21 crc kubenswrapper[4824]: I1124 13:16:21.113731 4824 generic.go:334] "Generic (PLEG): container finished" podID="2599a7bb-fad5-4c40-901b-189dc6687928" containerID="91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b" exitCode=1 Nov 24 13:16:21 crc kubenswrapper[4824]: I1124 13:16:21.113770 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ssstr" event={"ID":"2599a7bb-fad5-4c40-901b-189dc6687928","Type":"ContainerDied","Data":"91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b"} Nov 24 13:16:21 crc kubenswrapper[4824]: I1124 13:16:21.113827 4824 scope.go:117] "RemoveContainer" containerID="42c01b38d634f98cd5af313069392aa42c40b9ac87e1b15c1feb8a884b26a32c" Nov 24 13:16:21 crc kubenswrapper[4824]: I1124 13:16:21.114465 4824 scope.go:117] "RemoveContainer" containerID="91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b" Nov 24 13:16:21 crc kubenswrapper[4824]: E1124 13:16:21.114734 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-ssstr_openshift-multus(2599a7bb-fad5-4c40-901b-189dc6687928)\"" pod="openshift-multus/multus-ssstr" podUID="2599a7bb-fad5-4c40-901b-189dc6687928" Nov 24 13:16:22 crc kubenswrapper[4824]: I1124 13:16:22.010046 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:22 crc kubenswrapper[4824]: I1124 13:16:22.010175 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:22 crc kubenswrapper[4824]: E1124 13:16:22.010436 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:22 crc kubenswrapper[4824]: I1124 13:16:22.010761 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:22 crc kubenswrapper[4824]: E1124 13:16:22.010861 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:22 crc kubenswrapper[4824]: E1124 13:16:22.011092 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:22 crc kubenswrapper[4824]: I1124 13:16:22.119233 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/1.log" Nov 24 13:16:23 crc kubenswrapper[4824]: I1124 13:16:23.009934 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:23 crc kubenswrapper[4824]: E1124 13:16:23.010547 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:24 crc kubenswrapper[4824]: I1124 13:16:24.009758 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:24 crc kubenswrapper[4824]: I1124 13:16:24.009939 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:24 crc kubenswrapper[4824]: E1124 13:16:24.009974 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:24 crc kubenswrapper[4824]: E1124 13:16:24.010132 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:24 crc kubenswrapper[4824]: I1124 13:16:24.009766 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:24 crc kubenswrapper[4824]: E1124 13:16:24.010247 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:24 crc kubenswrapper[4824]: E1124 13:16:24.997919 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:16:25 crc kubenswrapper[4824]: I1124 13:16:25.009710 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:25 crc kubenswrapper[4824]: E1124 13:16:25.009897 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:25 crc kubenswrapper[4824]: I1124 13:16:25.010671 4824 scope.go:117] "RemoveContainer" containerID="377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42" Nov 24 13:16:25 crc kubenswrapper[4824]: E1124 13:16:25.010902 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nk5qz_openshift-ovn-kubernetes(0f43932f-b763-4fea-9ed9-737078e6548d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" Nov 24 13:16:26 crc kubenswrapper[4824]: I1124 13:16:26.009477 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:26 crc kubenswrapper[4824]: I1124 13:16:26.009609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:26 crc kubenswrapper[4824]: I1124 13:16:26.009771 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:26 crc kubenswrapper[4824]: E1124 13:16:26.009657 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:26 crc kubenswrapper[4824]: E1124 13:16:26.009895 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:26 crc kubenswrapper[4824]: E1124 13:16:26.010023 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:27 crc kubenswrapper[4824]: I1124 13:16:27.009124 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:27 crc kubenswrapper[4824]: E1124 13:16:27.009297 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:28 crc kubenswrapper[4824]: I1124 13:16:28.009880 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:28 crc kubenswrapper[4824]: I1124 13:16:28.009983 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:28 crc kubenswrapper[4824]: E1124 13:16:28.010112 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:28 crc kubenswrapper[4824]: E1124 13:16:28.010201 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:28 crc kubenswrapper[4824]: I1124 13:16:28.010716 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:28 crc kubenswrapper[4824]: E1124 13:16:28.010968 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:29 crc kubenswrapper[4824]: I1124 13:16:29.009145 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:29 crc kubenswrapper[4824]: E1124 13:16:29.009327 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:29 crc kubenswrapper[4824]: E1124 13:16:29.999829 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:16:30 crc kubenswrapper[4824]: I1124 13:16:30.009304 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:30 crc kubenswrapper[4824]: E1124 13:16:30.009439 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:30 crc kubenswrapper[4824]: I1124 13:16:30.010021 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:30 crc kubenswrapper[4824]: E1124 13:16:30.010159 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:30 crc kubenswrapper[4824]: I1124 13:16:30.010211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:30 crc kubenswrapper[4824]: E1124 13:16:30.010376 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:31 crc kubenswrapper[4824]: I1124 13:16:31.009356 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:31 crc kubenswrapper[4824]: E1124 13:16:31.010460 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:32 crc kubenswrapper[4824]: I1124 13:16:32.009733 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:32 crc kubenswrapper[4824]: I1124 13:16:32.009743 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:32 crc kubenswrapper[4824]: I1124 13:16:32.010040 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:32 crc kubenswrapper[4824]: E1124 13:16:32.010237 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:32 crc kubenswrapper[4824]: E1124 13:16:32.010377 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:32 crc kubenswrapper[4824]: E1124 13:16:32.010496 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:33 crc kubenswrapper[4824]: I1124 13:16:33.010225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:33 crc kubenswrapper[4824]: E1124 13:16:33.010434 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:34 crc kubenswrapper[4824]: I1124 13:16:34.009787 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:34 crc kubenswrapper[4824]: I1124 13:16:34.009946 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:34 crc kubenswrapper[4824]: I1124 13:16:34.009965 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:34 crc kubenswrapper[4824]: E1124 13:16:34.010136 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:34 crc kubenswrapper[4824]: I1124 13:16:34.010428 4824 scope.go:117] "RemoveContainer" containerID="91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b" Nov 24 13:16:34 crc kubenswrapper[4824]: E1124 13:16:34.010469 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:34 crc kubenswrapper[4824]: E1124 13:16:34.010671 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:34 crc kubenswrapper[4824]: I1124 13:16:34.164445 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/1.log" Nov 24 13:16:35 crc kubenswrapper[4824]: E1124 13:16:35.001199 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:16:35 crc kubenswrapper[4824]: I1124 13:16:35.009612 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:35 crc kubenswrapper[4824]: E1124 13:16:35.009763 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:35 crc kubenswrapper[4824]: I1124 13:16:35.169623 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/1.log" Nov 24 13:16:35 crc kubenswrapper[4824]: I1124 13:16:35.169687 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ssstr" event={"ID":"2599a7bb-fad5-4c40-901b-189dc6687928","Type":"ContainerStarted","Data":"b79ac592a95c8705cff781ef8fb18056df813f18c8c2b01fe974b33aa5140e70"} Nov 24 13:16:36 crc kubenswrapper[4824]: I1124 13:16:36.009829 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:36 crc kubenswrapper[4824]: I1124 13:16:36.010153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:36 crc kubenswrapper[4824]: I1124 13:16:36.010274 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:36 crc kubenswrapper[4824]: E1124 13:16:36.010420 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:36 crc kubenswrapper[4824]: E1124 13:16:36.010525 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:36 crc kubenswrapper[4824]: E1124 13:16:36.010582 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:37 crc kubenswrapper[4824]: I1124 13:16:37.010093 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:37 crc kubenswrapper[4824]: E1124 13:16:37.010232 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.009613 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.009679 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.009760 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:38 crc kubenswrapper[4824]: E1124 13:16:38.009757 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:38 crc kubenswrapper[4824]: E1124 13:16:38.010104 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:38 crc kubenswrapper[4824]: E1124 13:16:38.010284 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.010377 4824 scope.go:117] "RemoveContainer" containerID="377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42" Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.183313 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/3.log" Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.185944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerStarted","Data":"7e05e77c18a194c91bfe8d606ea0554e093dee70d4d5fd38ed73e5886750c7db"} Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.186380 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.214848 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podStartSLOduration=113.214830334 podStartE2EDuration="1m53.214830334s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:38.212728241 +0000 UTC m=+139.852267561" watchObservedRunningTime="2025-11-24 13:16:38.214830334 +0000 UTC m=+139.854369644" Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.945432 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-sc887"] Nov 24 13:16:38 crc kubenswrapper[4824]: I1124 13:16:38.945532 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:38 crc kubenswrapper[4824]: E1124 13:16:38.945625 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:39 crc kubenswrapper[4824]: I1124 13:16:39.009547 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:39 crc kubenswrapper[4824]: E1124 13:16:39.010726 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:40 crc kubenswrapper[4824]: E1124 13:16:40.003169 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:16:40 crc kubenswrapper[4824]: I1124 13:16:40.009560 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:40 crc kubenswrapper[4824]: E1124 13:16:40.009752 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:40 crc kubenswrapper[4824]: I1124 13:16:40.009571 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:40 crc kubenswrapper[4824]: E1124 13:16:40.010352 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:41 crc kubenswrapper[4824]: I1124 13:16:41.009881 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:41 crc kubenswrapper[4824]: E1124 13:16:41.010077 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:41 crc kubenswrapper[4824]: I1124 13:16:41.010431 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:41 crc kubenswrapper[4824]: E1124 13:16:41.010552 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:42 crc kubenswrapper[4824]: I1124 13:16:42.009289 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:42 crc kubenswrapper[4824]: I1124 13:16:42.009398 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:42 crc kubenswrapper[4824]: E1124 13:16:42.009509 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:42 crc kubenswrapper[4824]: E1124 13:16:42.009649 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:43 crc kubenswrapper[4824]: I1124 13:16:43.010052 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:43 crc kubenswrapper[4824]: I1124 13:16:43.010110 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:43 crc kubenswrapper[4824]: E1124 13:16:43.010213 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:16:43 crc kubenswrapper[4824]: E1124 13:16:43.010353 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sc887" podUID="7e81fafe-6567-47f5-bb87-78fffb25fa9b" Nov 24 13:16:44 crc kubenswrapper[4824]: I1124 13:16:44.009751 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:44 crc kubenswrapper[4824]: I1124 13:16:44.009914 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:44 crc kubenswrapper[4824]: E1124 13:16:44.010089 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:16:44 crc kubenswrapper[4824]: E1124 13:16:44.009927 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:16:45 crc kubenswrapper[4824]: I1124 13:16:45.009345 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:16:45 crc kubenswrapper[4824]: I1124 13:16:45.009670 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:45 crc kubenswrapper[4824]: I1124 13:16:45.011760 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 13:16:45 crc kubenswrapper[4824]: I1124 13:16:45.012036 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 13:16:45 crc kubenswrapper[4824]: I1124 13:16:45.012261 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 13:16:45 crc kubenswrapper[4824]: I1124 13:16:45.015937 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 13:16:46 crc kubenswrapper[4824]: I1124 13:16:46.009079 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:46 crc kubenswrapper[4824]: I1124 13:16:46.009079 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:46 crc kubenswrapper[4824]: I1124 13:16:46.011505 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 13:16:46 crc kubenswrapper[4824]: I1124 13:16:46.011763 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 13:16:47 crc kubenswrapper[4824]: I1124 13:16:47.936368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:47 crc kubenswrapper[4824]: E1124 13:16:47.936551 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:18:49.936523881 +0000 UTC m=+271.576063191 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:47 crc kubenswrapper[4824]: I1124 13:16:47.936653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:47 crc kubenswrapper[4824]: I1124 13:16:47.936707 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:47 crc kubenswrapper[4824]: I1124 13:16:47.942395 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:47 crc kubenswrapper[4824]: I1124 13:16:47.947025 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:48 crc kubenswrapper[4824]: I1124 13:16:48.038095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:48 crc kubenswrapper[4824]: I1124 13:16:48.038164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:48 crc kubenswrapper[4824]: I1124 13:16:48.039347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:48 crc kubenswrapper[4824]: I1124 13:16:48.041760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:48 crc kubenswrapper[4824]: I1124 13:16:48.047117 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:16:48 crc kubenswrapper[4824]: I1124 13:16:48.122248 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:48 crc kubenswrapper[4824]: I1124 13:16:48.129843 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:16:48 crc kubenswrapper[4824]: W1124 13:16:48.343262 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-f171f1d9c6c6d60df2671ddd9954d1c24c7d43e8c985728ce2df8b0d31c9e48e WatchSource:0}: Error finding container f171f1d9c6c6d60df2671ddd9954d1c24c7d43e8c985728ce2df8b0d31c9e48e: Status 404 returned error can't find the container with id f171f1d9c6c6d60df2671ddd9954d1c24c7d43e8c985728ce2df8b0d31c9e48e Nov 24 13:16:48 crc kubenswrapper[4824]: W1124 13:16:48.370548 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-3206c047a41487b0df7996a5bdfe654a98feec2086a3bccaf74ec13f6cb327d6 WatchSource:0}: Error finding container 3206c047a41487b0df7996a5bdfe654a98feec2086a3bccaf74ec13f6cb327d6: Status 404 returned error can't find the container with id 3206c047a41487b0df7996a5bdfe654a98feec2086a3bccaf74ec13f6cb327d6 Nov 24 13:16:49 crc kubenswrapper[4824]: I1124 13:16:49.226062 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1333fac49b5ff9e1341fdb2054950dfb90910b010d249220237f7a6ec5b43d10"} Nov 24 13:16:49 crc kubenswrapper[4824]: I1124 13:16:49.226129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3206c047a41487b0df7996a5bdfe654a98feec2086a3bccaf74ec13f6cb327d6"} Nov 24 13:16:49 crc kubenswrapper[4824]: I1124 13:16:49.226314 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:16:49 crc kubenswrapper[4824]: I1124 13:16:49.227177 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2192ff65861249f19534a22c84956d2eddfdf5bc0ad66a2823a3a4dc2cbaa8d4"} Nov 24 13:16:49 crc kubenswrapper[4824]: I1124 13:16:49.227197 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"42b97b61be6a1df358c912b32211d2341305e3e1aaed223519a1e0d94c3a4143"} Nov 24 13:16:49 crc kubenswrapper[4824]: I1124 13:16:49.228379 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fbc8e760be67f1308d815a49e874956676b23e3e1bb81ed38ffe89fb43d20da3"} Nov 24 13:16:49 crc kubenswrapper[4824]: I1124 13:16:49.228404 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f171f1d9c6c6d60df2671ddd9954d1c24c7d43e8c985728ce2df8b0d31c9e48e"} Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.495967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.536901 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.537383 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.541406 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.542518 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.542849 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lm6d6"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.542878 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.543910 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.545988 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.556342 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.556773 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.558479 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p5r76"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.559662 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.561395 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.562367 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.564984 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t2fdq"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.565682 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.568320 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.569753 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.569890 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.569994 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.570004 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.570100 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.571249 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.571925 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.575542 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.575735 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.576001 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.576122 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.576280 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.576400 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.577745 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5p99t"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.578350 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.583468 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.583654 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.583740 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.583740 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.584919 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.585075 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.585232 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.585633 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.585920 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.588636 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.588699 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.589499 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.590234 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.591542 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.591560 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.591705 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.592904 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.593128 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.593355 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.593403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.593491 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.593608 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.594219 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.595306 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.595739 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.595901 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.596084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.596197 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.596330 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.597132 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.597205 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.599118 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.599264 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.599415 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.599715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.599864 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-p6x6w"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.600313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.608614 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8wzb5"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.608874 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.609045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.615224 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.615424 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.616460 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.617195 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.617830 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.618124 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.618203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.618304 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.619023 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.619131 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.621086 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.621345 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.621453 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.621589 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.621926 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.622170 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.622194 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.622332 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.626926 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.627460 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.628556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.628697 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.628740 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.628979 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.643045 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.643563 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.644487 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dw5hv"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.673225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.674548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.678802 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-config\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.678862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.678895 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-image-import-ca\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.678921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60b8f545-2446-4422-8838-ac6b54ea119f-serving-cert\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.678947 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64d54af6-d194-48c1-9630-58d50c690eb3-serving-cert\") pod \"openshift-config-operator-7777fb866f-zp9xk\" (UID: \"64d54af6-d194-48c1-9630-58d50c690eb3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.678968 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/60b8f545-2446-4422-8838-ac6b54ea119f-node-pullsecrets\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.678991 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv9rq\" (UniqueName: \"kubernetes.io/projected/c974d46f-109b-4f2d-beb1-0f9ddae30b48-kube-api-access-qv9rq\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.679013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5efdc9d2-2541-4550-9f68-485fbc54473f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.679036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.679056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5efdc9d2-2541-4550-9f68-485fbc54473f-etcd-client\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.679082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-audit\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.679594 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.681329 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682590 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682722 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682755 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6t44\" (UniqueName: \"kubernetes.io/projected/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-kube-api-access-m6t44\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682799 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9wvm\" (UniqueName: \"kubernetes.io/projected/64d54af6-d194-48c1-9630-58d50c690eb3-kube-api-access-q9wvm\") pod \"openshift-config-operator-7777fb866f-zp9xk\" (UID: \"64d54af6-d194-48c1-9630-58d50c690eb3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/831ebed0-2a1f-490a-b315-27f535e22fe0-serving-cert\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682872 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dsn7\" (UniqueName: \"kubernetes.io/projected/fcf8e28b-55ce-4fd8-891b-50e2381d29a4-kube-api-access-2dsn7\") pod \"cluster-samples-operator-665b6dd947-vd2g5\" (UID: \"fcf8e28b-55ce-4fd8-891b-50e2381d29a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682942 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.682971 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-etcd-serving-ca\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683025 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjww5\" (UniqueName: \"kubernetes.io/projected/60b8f545-2446-4422-8838-ac6b54ea119f-kube-api-access-cjww5\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683049 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683077 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9ftj\" (UniqueName: \"kubernetes.io/projected/8a59be8e-4194-4e1c-89bd-3136366ed13b-kube-api-access-n9ftj\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxwck\" (UniqueName: \"kubernetes.io/projected/be5b6bdb-7ccd-4401-a46e-4c67fcd1c586-kube-api-access-vxwck\") pod \"openshift-apiserver-operator-796bbdcf4f-6s7gg\" (UID: \"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be5b6bdb-7ccd-4401-a46e-4c67fcd1c586-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6s7gg\" (UID: \"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5efdc9d2-2541-4550-9f68-485fbc54473f-serving-cert\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683177 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683203 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a59be8e-4194-4e1c-89bd-3136366ed13b-config\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683230 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-client-ca\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683274 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-client-ca\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683298 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683325 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5efdc9d2-2541-4550-9f68-485fbc54473f-encryption-config\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683355 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60b8f545-2446-4422-8838-ac6b54ea119f-etcd-client\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-service-ca-bundle\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683404 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72667\" (UniqueName: \"kubernetes.io/projected/831ebed0-2a1f-490a-b315-27f535e22fe0-kube-api-access-72667\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683446 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93a94b2f-1727-44d6-8438-c1492485cf06-serving-cert\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8a59be8e-4194-4e1c-89bd-3136366ed13b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fcf8e28b-55ce-4fd8-891b-50e2381d29a4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vd2g5\" (UID: \"fcf8e28b-55ce-4fd8-891b-50e2381d29a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683518 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683569 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c974d46f-109b-4f2d-beb1-0f9ddae30b48-config\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683591 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-dir\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683615 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60b8f545-2446-4422-8838-ac6b54ea119f-audit-dir\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26l8s\" (UniqueName: \"kubernetes.io/projected/f4875b41-d0f2-48c9-a451-839cd764a9cb-kube-api-access-26l8s\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9swp\" (UniqueName: \"kubernetes.io/projected/93a94b2f-1727-44d6-8438-c1492485cf06-kube-api-access-x9swp\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-policies\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683708 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-config\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5efdc9d2-2541-4550-9f68-485fbc54473f-audit-dir\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/64d54af6-d194-48c1-9630-58d50c690eb3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zp9xk\" (UID: \"64d54af6-d194-48c1-9630-58d50c690eb3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683786 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be5b6bdb-7ccd-4401-a46e-4c67fcd1c586-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6s7gg\" (UID: \"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92cqh\" (UniqueName: \"kubernetes.io/projected/5efdc9d2-2541-4550-9f68-485fbc54473f-kube-api-access-92cqh\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683887 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c974d46f-109b-4f2d-beb1-0f9ddae30b48-trusted-ca\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/60b8f545-2446-4422-8838-ac6b54ea119f-encryption-config\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683931 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683954 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5efdc9d2-2541-4550-9f68-485fbc54473f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.683979 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c974d46f-109b-4f2d-beb1-0f9ddae30b48-serving-cert\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.684001 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-config\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.684028 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-config\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.684050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-serving-cert\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.684072 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.684100 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8a59be8e-4194-4e1c-89bd-3136366ed13b-images\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.684122 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5efdc9d2-2541-4550-9f68-485fbc54473f-audit-policies\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.686022 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ll2hq"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.686408 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vwk9h"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.686621 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.686916 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.687207 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.687828 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.688145 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vwk9h" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.694609 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.695448 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.696849 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.697025 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.697187 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.697277 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.710378 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.710581 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.710898 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.711111 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.711211 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.711311 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.711389 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.711460 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.711646 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.711780 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.711928 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.712134 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.712240 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.712332 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.712439 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.712548 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.712951 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zrmh8"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.713388 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.713906 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.714191 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.721541 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.721631 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.721853 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.724150 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.724709 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.730000 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j9k4z"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.730799 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.733293 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ps582"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.733973 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.740859 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.741581 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.741965 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.743165 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-7lwbq"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.744546 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.764066 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.766352 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.772558 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.775292 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.776743 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.785699 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.786031 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.793788 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpfnz\" (UniqueName: \"kubernetes.io/projected/2cfbf6ab-805e-4776-8f3e-1823b675c1be-kube-api-access-vpfnz\") pod \"dns-operator-744455d44c-dw5hv\" (UID: \"2cfbf6ab-805e-4776-8f3e-1823b675c1be\") " pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.793861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c974d46f-109b-4f2d-beb1-0f9ddae30b48-trusted-ca\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.793889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/60b8f545-2446-4422-8838-ac6b54ea119f-encryption-config\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.793914 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.793936 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5efdc9d2-2541-4550-9f68-485fbc54473f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.793959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c974d46f-109b-4f2d-beb1-0f9ddae30b48-serving-cert\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.793980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-config\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794005 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-config\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794026 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-serving-cert\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8a59be8e-4194-4e1c-89bd-3136366ed13b-images\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794100 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlhgs\" (UniqueName: \"kubernetes.io/projected/269ad190-1f61-4290-813b-86950552ea1f-kube-api-access-wlhgs\") pod \"openshift-controller-manager-operator-756b6f6bc6-t9jw7\" (UID: \"269ad190-1f61-4290-813b-86950552ea1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794124 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5efdc9d2-2541-4550-9f68-485fbc54473f-audit-policies\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794146 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-config\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/774e51b9-8e46-458f-82d2-e70027a8a0fc-auth-proxy-config\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-image-import-ca\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-config\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794256 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60b8f545-2446-4422-8838-ac6b54ea119f-serving-cert\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64d54af6-d194-48c1-9630-58d50c690eb3-serving-cert\") pod \"openshift-config-operator-7777fb866f-zp9xk\" (UID: \"64d54af6-d194-48c1-9630-58d50c690eb3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e7668d3-1c2b-4c90-9940-c1898a6c5069-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vxl9p\" (UID: \"7e7668d3-1c2b-4c90-9940-c1898a6c5069\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/60b8f545-2446-4422-8838-ac6b54ea119f-node-pullsecrets\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv9rq\" (UniqueName: \"kubernetes.io/projected/c974d46f-109b-4f2d-beb1-0f9ddae30b48-kube-api-access-qv9rq\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5efdc9d2-2541-4550-9f68-485fbc54473f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794383 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794408 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5efdc9d2-2541-4550-9f68-485fbc54473f-etcd-client\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qt47\" (UniqueName: \"kubernetes.io/projected/774e51b9-8e46-458f-82d2-e70027a8a0fc-kube-api-access-6qt47\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794454 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9wvm\" (UniqueName: \"kubernetes.io/projected/64d54af6-d194-48c1-9630-58d50c690eb3-kube-api-access-q9wvm\") pod \"openshift-config-operator-7777fb866f-zp9xk\" (UID: \"64d54af6-d194-48c1-9630-58d50c690eb3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794476 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-audit\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794496 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6t44\" (UniqueName: \"kubernetes.io/projected/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-kube-api-access-m6t44\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794518 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/831ebed0-2a1f-490a-b315-27f535e22fe0-serving-cert\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794583 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dsn7\" (UniqueName: \"kubernetes.io/projected/fcf8e28b-55ce-4fd8-891b-50e2381d29a4-kube-api-access-2dsn7\") pod \"cluster-samples-operator-665b6dd947-vd2g5\" (UID: \"fcf8e28b-55ce-4fd8-891b-50e2381d29a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794603 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794626 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-etcd-serving-ca\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794674 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjww5\" (UniqueName: \"kubernetes.io/projected/60b8f545-2446-4422-8838-ac6b54ea119f-kube-api-access-cjww5\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9ftj\" (UniqueName: \"kubernetes.io/projected/8a59be8e-4194-4e1c-89bd-3136366ed13b-kube-api-access-n9ftj\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxwck\" (UniqueName: \"kubernetes.io/projected/be5b6bdb-7ccd-4401-a46e-4c67fcd1c586-kube-api-access-vxwck\") pod \"openshift-apiserver-operator-796bbdcf4f-6s7gg\" (UID: \"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-service-ca\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794793 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be5b6bdb-7ccd-4401-a46e-4c67fcd1c586-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6s7gg\" (UID: \"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5efdc9d2-2541-4550-9f68-485fbc54473f-serving-cert\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794860 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794886 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-oauth-config\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794907 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/774e51b9-8e46-458f-82d2-e70027a8a0fc-machine-approver-tls\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794929 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a59be8e-4194-4e1c-89bd-3136366ed13b-config\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqsxl\" (UniqueName: \"kubernetes.io/projected/76c0d257-aca8-434e-97b2-067ddc1e7f66-kube-api-access-dqsxl\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.794978 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6s5v\" (UniqueName: \"kubernetes.io/projected/ecea50ae-16db-46d7-a9cd-84a339185ae4-kube-api-access-v6s5v\") pod \"downloads-7954f5f757-vwk9h\" (UID: \"ecea50ae-16db-46d7-a9cd-84a339185ae4\") " pod="openshift-console/downloads-7954f5f757-vwk9h" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-client-ca\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795024 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-client-ca\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795047 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5efdc9d2-2541-4550-9f68-485fbc54473f-encryption-config\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795093 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/269ad190-1f61-4290-813b-86950552ea1f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-t9jw7\" (UID: \"269ad190-1f61-4290-813b-86950552ea1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795116 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60b8f545-2446-4422-8838-ac6b54ea119f-etcd-client\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-service-ca-bundle\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795162 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72667\" (UniqueName: \"kubernetes.io/projected/831ebed0-2a1f-490a-b315-27f535e22fe0-kube-api-access-72667\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fcf8e28b-55ce-4fd8-891b-50e2381d29a4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vd2g5\" (UID: \"fcf8e28b-55ce-4fd8-891b-50e2381d29a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93a94b2f-1727-44d6-8438-c1492485cf06-serving-cert\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795244 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8a59be8e-4194-4e1c-89bd-3136366ed13b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795291 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-trusted-ca-bundle\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7668d3-1c2b-4c90-9940-c1898a6c5069-config\") pod \"kube-apiserver-operator-766d6c64bb-vxl9p\" (UID: \"7e7668d3-1c2b-4c90-9940-c1898a6c5069\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795351 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-oauth-serving-cert\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795373 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c974d46f-109b-4f2d-beb1-0f9ddae30b48-config\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-dir\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795413 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7668d3-1c2b-4c90-9940-c1898a6c5069-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vxl9p\" (UID: \"7e7668d3-1c2b-4c90-9940-c1898a6c5069\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795436 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60b8f545-2446-4422-8838-ac6b54ea119f-audit-dir\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795457 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26l8s\" (UniqueName: \"kubernetes.io/projected/f4875b41-d0f2-48c9-a451-839cd764a9cb-kube-api-access-26l8s\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-policies\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/774e51b9-8e46-458f-82d2-e70027a8a0fc-config\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2cfbf6ab-805e-4776-8f3e-1823b675c1be-metrics-tls\") pod \"dns-operator-744455d44c-dw5hv\" (UID: \"2cfbf6ab-805e-4776-8f3e-1823b675c1be\") " pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795569 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269ad190-1f61-4290-813b-86950552ea1f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-t9jw7\" (UID: \"269ad190-1f61-4290-813b-86950552ea1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9swp\" (UniqueName: \"kubernetes.io/projected/93a94b2f-1727-44d6-8438-c1492485cf06-kube-api-access-x9swp\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795608 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795943 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.796199 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p5r76"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.796212 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.796467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-config\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.796550 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.798064 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.798825 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.799818 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.800153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.795619 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-config\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.800349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5efdc9d2-2541-4550-9f68-485fbc54473f-audit-dir\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.800370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/64d54af6-d194-48c1-9630-58d50c690eb3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zp9xk\" (UID: \"64d54af6-d194-48c1-9630-58d50c690eb3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.800400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92cqh\" (UniqueName: \"kubernetes.io/projected/5efdc9d2-2541-4550-9f68-485fbc54473f-kube-api-access-92cqh\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.800419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.800442 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-serving-cert\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.800465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be5b6bdb-7ccd-4401-a46e-4c67fcd1c586-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6s7gg\" (UID: \"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.802822 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.804818 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c974d46f-109b-4f2d-beb1-0f9ddae30b48-trusted-ca\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.805115 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/60b8f545-2446-4422-8838-ac6b54ea119f-node-pullsecrets\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.805859 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5efdc9d2-2541-4550-9f68-485fbc54473f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.806410 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.807298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.807970 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.808181 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-dir\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.809467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.809829 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.809849 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be5b6bdb-7ccd-4401-a46e-4c67fcd1c586-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6s7gg\" (UID: \"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.812506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60b8f545-2446-4422-8838-ac6b54ea119f-audit-dir\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.813465 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-policies\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.814290 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-etcd-serving-ca\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.815036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/64d54af6-d194-48c1-9630-58d50c690eb3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zp9xk\" (UID: \"64d54af6-d194-48c1-9630-58d50c690eb3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.815938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5efdc9d2-2541-4550-9f68-485fbc54473f-audit-dir\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.817994 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-audit\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.818700 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c974d46f-109b-4f2d-beb1-0f9ddae30b48-config\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.819656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-client-ca\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.821557 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.822429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be5b6bdb-7ccd-4401-a46e-4c67fcd1c586-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6s7gg\" (UID: \"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.822680 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.823684 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ftftq"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.824410 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.827619 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mzl5h"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.828341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.828542 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.829240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a59be8e-4194-4e1c-89bd-3136366ed13b-config\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.830089 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.830218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5efdc9d2-2541-4550-9f68-485fbc54473f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.830583 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.830903 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.831348 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-service-ca-bundle\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.831383 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4fpg6"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.831889 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.832196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-config\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.832270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.832247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.832714 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93a94b2f-1727-44d6-8438-c1492485cf06-serving-cert\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.832733 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zq9w9"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.833101 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c974d46f-109b-4f2d-beb1-0f9ddae30b48-serving-cert\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.833907 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.834725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8a59be8e-4194-4e1c-89bd-3136366ed13b-images\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.835263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5efdc9d2-2541-4550-9f68-485fbc54473f-audit-policies\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.835493 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.836134 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-config\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.836600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/60b8f545-2446-4422-8838-ac6b54ea119f-image-import-ca\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.836768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-client-ca\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.836777 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-config\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.836472 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60b8f545-2446-4422-8838-ac6b54ea119f-etcd-client\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.837728 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-serving-cert\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.839088 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/60b8f545-2446-4422-8838-ac6b54ea119f-encryption-config\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.839145 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lm6d6"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.840992 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.841713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60b8f545-2446-4422-8838-ac6b54ea119f-serving-cert\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.841843 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.843307 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t2fdq"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.843403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.843953 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.846154 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.846247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.849923 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.850798 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.851636 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.853759 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5efdc9d2-2541-4550-9f68-485fbc54473f-serving-cert\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.853958 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5p99t"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.854067 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.854095 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.855134 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.855349 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64d54af6-d194-48c1-9630-58d50c690eb3-serving-cert\") pod \"openshift-config-operator-7777fb866f-zp9xk\" (UID: \"64d54af6-d194-48c1-9630-58d50c690eb3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.856219 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vmp98"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.857381 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.857438 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/831ebed0-2a1f-490a-b315-27f535e22fe0-serving-cert\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.858122 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5efdc9d2-2541-4550-9f68-485fbc54473f-encryption-config\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.858670 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5efdc9d2-2541-4550-9f68-485fbc54473f-etcd-client\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.858686 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fcf8e28b-55ce-4fd8-891b-50e2381d29a4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vd2g5\" (UID: \"fcf8e28b-55ce-4fd8-891b-50e2381d29a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.858755 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-p6x6w"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.859030 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.859316 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8a59be8e-4194-4e1c-89bd-3136366ed13b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.860859 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ll2hq"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.860902 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8wzb5"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.861173 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.862727 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.862770 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.863837 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.864599 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.865755 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dw5hv"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.868704 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j9k4z"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.868737 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vwk9h"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.868749 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-fnqzp"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.869250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fnqzp" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.873348 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ksxcs"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.874374 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ps582"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.874400 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.874483 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.875657 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.879380 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.880390 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.881047 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.882875 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.891010 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zrmh8"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.891083 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.898277 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.900987 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901160 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7668d3-1c2b-4c90-9940-c1898a6c5069-config\") pod \"kube-apiserver-operator-766d6c64bb-vxl9p\" (UID: \"7e7668d3-1c2b-4c90-9940-c1898a6c5069\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901200 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-oauth-serving-cert\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/adc0b256-e903-4e81-8d37-8f44d27234d1-etcd-ca\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7668d3-1c2b-4c90-9940-c1898a6c5069-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vxl9p\" (UID: \"7e7668d3-1c2b-4c90-9940-c1898a6c5069\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269ad190-1f61-4290-813b-86950552ea1f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-t9jw7\" (UID: \"269ad190-1f61-4290-813b-86950552ea1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/774e51b9-8e46-458f-82d2-e70027a8a0fc-config\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2cfbf6ab-805e-4776-8f3e-1823b675c1be-metrics-tls\") pod \"dns-operator-744455d44c-dw5hv\" (UID: \"2cfbf6ab-805e-4776-8f3e-1823b675c1be\") " pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc0b256-e903-4e81-8d37-8f44d27234d1-config\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-serving-cert\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901397 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d63938dd-7154-4aac-9071-59a7803032c9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ps582\" (UID: \"d63938dd-7154-4aac-9071-59a7803032c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901417 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpfnz\" (UniqueName: \"kubernetes.io/projected/2cfbf6ab-805e-4776-8f3e-1823b675c1be-kube-api-access-vpfnz\") pod \"dns-operator-744455d44c-dw5hv\" (UID: \"2cfbf6ab-805e-4776-8f3e-1823b675c1be\") " pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrrjn\" (UniqueName: \"kubernetes.io/projected/adc0b256-e903-4e81-8d37-8f44d27234d1-kube-api-access-lrrjn\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901463 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlhgs\" (UniqueName: \"kubernetes.io/projected/269ad190-1f61-4290-813b-86950552ea1f-kube-api-access-wlhgs\") pod \"openshift-controller-manager-operator-756b6f6bc6-t9jw7\" (UID: \"269ad190-1f61-4290-813b-86950552ea1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901484 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01ec77ac-5384-4457-8d8a-9d14135c2b82-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/090539a8-ec0a-4406-a781-2d45333291a1-images\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901521 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-config\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/774e51b9-8e46-458f-82d2-e70027a8a0fc-auth-proxy-config\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d63938dd-7154-4aac-9071-59a7803032c9-proxy-tls\") pod \"machine-config-controller-84d6567774-ps582\" (UID: \"d63938dd-7154-4aac-9071-59a7803032c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e7668d3-1c2b-4c90-9940-c1898a6c5069-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vxl9p\" (UID: \"7e7668d3-1c2b-4c90-9940-c1898a6c5069\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901628 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh2xb\" (UniqueName: \"kubernetes.io/projected/01ec77ac-5384-4457-8d8a-9d14135c2b82-kube-api-access-lh2xb\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901648 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8xxk\" (UniqueName: \"kubernetes.io/projected/d63938dd-7154-4aac-9071-59a7803032c9-kube-api-access-x8xxk\") pod \"machine-config-controller-84d6567774-ps582\" (UID: \"d63938dd-7154-4aac-9071-59a7803032c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/adc0b256-e903-4e81-8d37-8f44d27234d1-etcd-client\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901701 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qt47\" (UniqueName: \"kubernetes.io/projected/774e51b9-8e46-458f-82d2-e70027a8a0fc-kube-api-access-6qt47\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901724 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01ec77ac-5384-4457-8d8a-9d14135c2b82-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901744 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/090539a8-ec0a-4406-a781-2d45333291a1-proxy-tls\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901799 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/adc0b256-e903-4e81-8d37-8f44d27234d1-etcd-service-ca\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901842 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-service-ca\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xff6\" (UniqueName: \"kubernetes.io/projected/090539a8-ec0a-4406-a781-2d45333291a1-kube-api-access-4xff6\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-oauth-config\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/774e51b9-8e46-458f-82d2-e70027a8a0fc-machine-approver-tls\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/090539a8-ec0a-4406-a781-2d45333291a1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/01ec77ac-5384-4457-8d8a-9d14135c2b82-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901976 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqsxl\" (UniqueName: \"kubernetes.io/projected/76c0d257-aca8-434e-97b2-067ddc1e7f66-kube-api-access-dqsxl\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.901996 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6s5v\" (UniqueName: \"kubernetes.io/projected/ecea50ae-16db-46d7-a9cd-84a339185ae4-kube-api-access-v6s5v\") pod \"downloads-7954f5f757-vwk9h\" (UID: \"ecea50ae-16db-46d7-a9cd-84a339185ae4\") " pod="openshift-console/downloads-7954f5f757-vwk9h" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.902012 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/269ad190-1f61-4290-813b-86950552ea1f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-t9jw7\" (UID: \"269ad190-1f61-4290-813b-86950552ea1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.902031 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adc0b256-e903-4e81-8d37-8f44d27234d1-serving-cert\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.902059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-trusted-ca-bundle\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.902164 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7668d3-1c2b-4c90-9940-c1898a6c5069-config\") pod \"kube-apiserver-operator-766d6c64bb-vxl9p\" (UID: \"7e7668d3-1c2b-4c90-9940-c1898a6c5069\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.902243 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-oauth-serving-cert\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.902950 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269ad190-1f61-4290-813b-86950552ea1f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-t9jw7\" (UID: \"269ad190-1f61-4290-813b-86950552ea1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.903208 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-trusted-ca-bundle\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.903577 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/774e51b9-8e46-458f-82d2-e70027a8a0fc-config\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.906735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7668d3-1c2b-4c90-9940-c1898a6c5069-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vxl9p\" (UID: \"7e7668d3-1c2b-4c90-9940-c1898a6c5069\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.907124 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zq9w9"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.907439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-service-ca\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.908579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-config\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.909597 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/774e51b9-8e46-458f-82d2-e70027a8a0fc-auth-proxy-config\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.911663 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ftftq"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.912248 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/269ad190-1f61-4290-813b-86950552ea1f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-t9jw7\" (UID: \"269ad190-1f61-4290-813b-86950552ea1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.912678 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.913518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/774e51b9-8e46-458f-82d2-e70027a8a0fc-machine-approver-tls\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.914627 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-oauth-config\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.915456 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-serving-cert\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.916324 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.918671 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.921015 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.922341 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mzl5h"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.924548 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.926785 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vmp98"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.928125 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fnqzp"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.929630 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ksxcs"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.931011 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-7dczj"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.931688 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.932467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2cfbf6ab-805e-4776-8f3e-1823b675c1be-metrics-tls\") pod \"dns-operator-744455d44c-dw5hv\" (UID: \"2cfbf6ab-805e-4776-8f3e-1823b675c1be\") " pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.933751 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4fpg6"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.934916 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q"] Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.944977 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.966324 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.979702 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 13:16:50 crc kubenswrapper[4824]: I1124 13:16:50.999348 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003585 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adc0b256-e903-4e81-8d37-8f44d27234d1-serving-cert\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003683 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/adc0b256-e903-4e81-8d37-8f44d27234d1-etcd-ca\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003743 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc0b256-e903-4e81-8d37-8f44d27234d1-config\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003773 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d63938dd-7154-4aac-9071-59a7803032c9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ps582\" (UID: \"d63938dd-7154-4aac-9071-59a7803032c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003821 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrrjn\" (UniqueName: \"kubernetes.io/projected/adc0b256-e903-4e81-8d37-8f44d27234d1-kube-api-access-lrrjn\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003860 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01ec77ac-5384-4457-8d8a-9d14135c2b82-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/090539a8-ec0a-4406-a781-2d45333291a1-images\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003914 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d63938dd-7154-4aac-9071-59a7803032c9-proxy-tls\") pod \"machine-config-controller-84d6567774-ps582\" (UID: \"d63938dd-7154-4aac-9071-59a7803032c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh2xb\" (UniqueName: \"kubernetes.io/projected/01ec77ac-5384-4457-8d8a-9d14135c2b82-kube-api-access-lh2xb\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.003968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8xxk\" (UniqueName: \"kubernetes.io/projected/d63938dd-7154-4aac-9071-59a7803032c9-kube-api-access-x8xxk\") pod \"machine-config-controller-84d6567774-ps582\" (UID: \"d63938dd-7154-4aac-9071-59a7803032c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.004000 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/adc0b256-e903-4e81-8d37-8f44d27234d1-etcd-client\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.004046 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01ec77ac-5384-4457-8d8a-9d14135c2b82-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.004071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/090539a8-ec0a-4406-a781-2d45333291a1-proxy-tls\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.004139 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/adc0b256-e903-4e81-8d37-8f44d27234d1-etcd-service-ca\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.004165 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xff6\" (UniqueName: \"kubernetes.io/projected/090539a8-ec0a-4406-a781-2d45333291a1-kube-api-access-4xff6\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.004210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/090539a8-ec0a-4406-a781-2d45333291a1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.004235 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/01ec77ac-5384-4457-8d8a-9d14135c2b82-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.004846 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/090539a8-ec0a-4406-a781-2d45333291a1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.004921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01ec77ac-5384-4457-8d8a-9d14135c2b82-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.005899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d63938dd-7154-4aac-9071-59a7803032c9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ps582\" (UID: \"d63938dd-7154-4aac-9071-59a7803032c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.019369 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.039506 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.059613 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.069073 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/01ec77ac-5384-4457-8d8a-9d14135c2b82-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.080662 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.085769 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/adc0b256-e903-4e81-8d37-8f44d27234d1-etcd-ca\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.100751 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.119760 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.126398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adc0b256-e903-4e81-8d37-8f44d27234d1-serving-cert\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.139663 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.146738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/adc0b256-e903-4e81-8d37-8f44d27234d1-etcd-client\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.159632 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.164963 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/adc0b256-e903-4e81-8d37-8f44d27234d1-etcd-service-ca\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.180208 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.199695 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.220169 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.225570 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adc0b256-e903-4e81-8d37-8f44d27234d1-config\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.241234 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.249032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d63938dd-7154-4aac-9071-59a7803032c9-proxy-tls\") pod \"machine-config-controller-84d6567774-ps582\" (UID: \"d63938dd-7154-4aac-9071-59a7803032c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.259735 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.279398 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.299346 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.320920 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.339393 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.360178 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.380291 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.401305 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.419978 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.440304 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.459308 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.481246 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.499917 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.519546 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.540933 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.559973 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.598742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dsn7\" (UniqueName: \"kubernetes.io/projected/fcf8e28b-55ce-4fd8-891b-50e2381d29a4-kube-api-access-2dsn7\") pod \"cluster-samples-operator-665b6dd947-vd2g5\" (UID: \"fcf8e28b-55ce-4fd8-891b-50e2381d29a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.620565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6t44\" (UniqueName: \"kubernetes.io/projected/e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1-kube-api-access-m6t44\") pod \"authentication-operator-69f744f599-8wzb5\" (UID: \"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.637075 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72667\" (UniqueName: \"kubernetes.io/projected/831ebed0-2a1f-490a-b315-27f535e22fe0-kube-api-access-72667\") pod \"route-controller-manager-6576b87f9c-f9ngl\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.642545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.660123 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.673206 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv9rq\" (UniqueName: \"kubernetes.io/projected/c974d46f-109b-4f2d-beb1-0f9ddae30b48-kube-api-access-qv9rq\") pod \"console-operator-58897d9998-p6x6w\" (UID: \"c974d46f-109b-4f2d-beb1-0f9ddae30b48\") " pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.694104 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9ftj\" (UniqueName: \"kubernetes.io/projected/8a59be8e-4194-4e1c-89bd-3136366ed13b-kube-api-access-n9ftj\") pod \"machine-api-operator-5694c8668f-p5r76\" (UID: \"8a59be8e-4194-4e1c-89bd-3136366ed13b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.717717 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.722987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxwck\" (UniqueName: \"kubernetes.io/projected/be5b6bdb-7ccd-4401-a46e-4c67fcd1c586-kube-api-access-vxwck\") pod \"openshift-apiserver-operator-796bbdcf4f-6s7gg\" (UID: \"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.740382 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.742837 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjww5\" (UniqueName: \"kubernetes.io/projected/60b8f545-2446-4422-8838-ac6b54ea119f-kube-api-access-cjww5\") pod \"apiserver-76f77b778f-t2fdq\" (UID: \"60b8f545-2446-4422-8838-ac6b54ea119f\") " pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.760181 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.779358 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.794992 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.798121 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.799557 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.807297 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.817838 4824 request.go:700] Waited for 1.007267882s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serving-cert&limit=500&resourceVersion=0 Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.819490 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.823979 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.839377 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.861212 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5"] Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.863195 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.881244 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.882620 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.899155 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.955741 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9swp\" (UniqueName: \"kubernetes.io/projected/93a94b2f-1727-44d6-8438-c1492485cf06-kube-api-access-x9swp\") pod \"controller-manager-879f6c89f-lm6d6\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.956915 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26l8s\" (UniqueName: \"kubernetes.io/projected/f4875b41-d0f2-48c9-a451-839cd764a9cb-kube-api-access-26l8s\") pod \"oauth-openshift-558db77b4-5p99t\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.984386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9wvm\" (UniqueName: \"kubernetes.io/projected/64d54af6-d194-48c1-9630-58d50c690eb3-kube-api-access-q9wvm\") pod \"openshift-config-operator-7777fb866f-zp9xk\" (UID: \"64d54af6-d194-48c1-9630-58d50c690eb3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:51 crc kubenswrapper[4824]: I1124 13:16:51.984729 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.005664 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 13:16:52 crc kubenswrapper[4824]: E1124 13:16:52.006061 4824 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Nov 24 13:16:52 crc kubenswrapper[4824]: E1124 13:16:52.006186 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/090539a8-ec0a-4406-a781-2d45333291a1-proxy-tls podName:090539a8-ec0a-4406-a781-2d45333291a1 nodeName:}" failed. No retries permitted until 2025-11-24 13:16:52.506161743 +0000 UTC m=+154.145701073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/090539a8-ec0a-4406-a781-2d45333291a1-proxy-tls") pod "machine-config-operator-74547568cd-xxpwx" (UID: "090539a8-ec0a-4406-a781-2d45333291a1") : failed to sync secret cache: timed out waiting for the condition Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.006651 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/090539a8-ec0a-4406-a781-2d45333291a1-images\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.026797 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-p6x6w"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.027929 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.039612 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 13:16:52 crc kubenswrapper[4824]: W1124 13:16:52.041182 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc974d46f_109b_4f2d_beb1_0f9ddae30b48.slice/crio-d637dffc1cd1a700d6cb6537e704dde70b7294dd07f0894fe80a154172a6de54 WatchSource:0}: Error finding container d637dffc1cd1a700d6cb6537e704dde70b7294dd07f0894fe80a154172a6de54: Status 404 returned error can't find the container with id d637dffc1cd1a700d6cb6537e704dde70b7294dd07f0894fe80a154172a6de54 Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.080150 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.087791 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.101797 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.120187 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.150697 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t2fdq"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.158475 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.159488 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92cqh\" (UniqueName: \"kubernetes.io/projected/5efdc9d2-2541-4550-9f68-485fbc54473f-kube-api-access-92cqh\") pod \"apiserver-7bbb656c7d-857cb\" (UID: \"5efdc9d2-2541-4550-9f68-485fbc54473f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.160743 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 13:16:52 crc kubenswrapper[4824]: W1124 13:16:52.169026 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod831ebed0_2a1f_490a_b315_27f535e22fe0.slice/crio-bf2e7794b8481bb005a8ca8651f7b55becf31fc6fca4729d53fd3cf908bde62c WatchSource:0}: Error finding container bf2e7794b8481bb005a8ca8651f7b55becf31fc6fca4729d53fd3cf908bde62c: Status 404 returned error can't find the container with id bf2e7794b8481bb005a8ca8651f7b55becf31fc6fca4729d53fd3cf908bde62c Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.180724 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.196111 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p5r76"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.204247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.207378 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.220424 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.237368 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.239342 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.255886 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" event={"ID":"fcf8e28b-55ce-4fd8-891b-50e2381d29a4","Type":"ContainerStarted","Data":"3b7fd73993afe4301268bdf527bf3ccb2eedeba7c9130253d879e9a1bc2e27e2"} Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.259381 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.266504 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" event={"ID":"8a59be8e-4194-4e1c-89bd-3136366ed13b","Type":"ContainerStarted","Data":"86c2b303c6894282fcdf36a26ceea5ec533afbd352a480235b8b63859a631a3c"} Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.269731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-p6x6w" event={"ID":"c974d46f-109b-4f2d-beb1-0f9ddae30b48","Type":"ContainerStarted","Data":"8f880802dd90d383c3d187d38731dad8179c6761df3b9eac4e35941cc54b56e1"} Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.269774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-p6x6w" event={"ID":"c974d46f-109b-4f2d-beb1-0f9ddae30b48","Type":"ContainerStarted","Data":"d637dffc1cd1a700d6cb6537e704dde70b7294dd07f0894fe80a154172a6de54"} Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.270049 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.272133 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-p6x6w container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.272172 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-p6x6w" podUID="c974d46f-109b-4f2d-beb1-0f9ddae30b48" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.274284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" event={"ID":"60b8f545-2446-4422-8838-ac6b54ea119f","Type":"ContainerStarted","Data":"fbb1f31026bf095eecdc1bc434a939314709ed034fd630792d968ace9e07c605"} Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.278941 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.280714 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" event={"ID":"831ebed0-2a1f-490a-b315-27f535e22fe0","Type":"ContainerStarted","Data":"bf2e7794b8481bb005a8ca8651f7b55becf31fc6fca4729d53fd3cf908bde62c"} Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.300937 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.321927 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.340760 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.366116 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.380225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.388437 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.395408 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lm6d6"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.400391 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 13:16:52 crc kubenswrapper[4824]: W1124 13:16:52.410424 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93a94b2f_1727_44d6_8438_c1492485cf06.slice/crio-770475f891428760a1ec46278bb77e6dc845deb090e432c79f48d86402599dab WatchSource:0}: Error finding container 770475f891428760a1ec46278bb77e6dc845deb090e432c79f48d86402599dab: Status 404 returned error can't find the container with id 770475f891428760a1ec46278bb77e6dc845deb090e432c79f48d86402599dab Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.423318 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.445397 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.459727 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.478156 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.485778 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.491844 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8wzb5"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.499426 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.519140 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.546778 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.549237 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5p99t"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.553792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/090539a8-ec0a-4406-a781-2d45333291a1-proxy-tls\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.561154 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.563755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/090539a8-ec0a-4406-a781-2d45333291a1-proxy-tls\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:52 crc kubenswrapper[4824]: W1124 13:16:52.573253 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7abfe4e_76f0_4a5a_9dcd_50a80dd8d3e1.slice/crio-cfd592a9038b08454a0c4effa9d4ac8695f67fa2e23a9d24ec9eb6e1c7493593 WatchSource:0}: Error finding container cfd592a9038b08454a0c4effa9d4ac8695f67fa2e23a9d24ec9eb6e1c7493593: Status 404 returned error can't find the container with id cfd592a9038b08454a0c4effa9d4ac8695f67fa2e23a9d24ec9eb6e1c7493593 Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.583054 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.609348 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.621836 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.640506 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.665381 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.681597 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.705518 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.723301 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.744523 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb"] Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.744638 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.765223 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.789338 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.800513 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.817990 4824 request.go:700] Waited for 1.943240062s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.824410 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.873706 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qt47\" (UniqueName: \"kubernetes.io/projected/774e51b9-8e46-458f-82d2-e70027a8a0fc-kube-api-access-6qt47\") pod \"machine-approver-56656f9798-rk5nj\" (UID: \"774e51b9-8e46-458f-82d2-e70027a8a0fc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.880327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6s5v\" (UniqueName: \"kubernetes.io/projected/ecea50ae-16db-46d7-a9cd-84a339185ae4-kube-api-access-v6s5v\") pod \"downloads-7954f5f757-vwk9h\" (UID: \"ecea50ae-16db-46d7-a9cd-84a339185ae4\") " pod="openshift-console/downloads-7954f5f757-vwk9h" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.899011 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqsxl\" (UniqueName: \"kubernetes.io/projected/76c0d257-aca8-434e-97b2-067ddc1e7f66-kube-api-access-dqsxl\") pod \"console-f9d7485db-ll2hq\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.919618 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlhgs\" (UniqueName: \"kubernetes.io/projected/269ad190-1f61-4290-813b-86950552ea1f-kube-api-access-wlhgs\") pod \"openshift-controller-manager-operator-756b6f6bc6-t9jw7\" (UID: \"269ad190-1f61-4290-813b-86950552ea1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.948552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpfnz\" (UniqueName: \"kubernetes.io/projected/2cfbf6ab-805e-4776-8f3e-1823b675c1be-kube-api-access-vpfnz\") pod \"dns-operator-744455d44c-dw5hv\" (UID: \"2cfbf6ab-805e-4776-8f3e-1823b675c1be\") " pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.958486 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e7668d3-1c2b-4c90-9940-c1898a6c5069-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vxl9p\" (UID: \"7e7668d3-1c2b-4c90-9940-c1898a6c5069\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.959559 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.979778 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 13:16:52 crc kubenswrapper[4824]: I1124 13:16:52.999399 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.005213 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.020793 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.029211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.033454 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01ec77ac-5384-4457-8d8a-9d14135c2b82-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.036215 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.043949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vwk9h" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.055110 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh2xb\" (UniqueName: \"kubernetes.io/projected/01ec77ac-5384-4457-8d8a-9d14135c2b82-kube-api-access-lh2xb\") pod \"cluster-image-registry-operator-dc59b4c8b-ss8j9\" (UID: \"01ec77ac-5384-4457-8d8a-9d14135c2b82\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.075065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.076196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8xxk\" (UniqueName: \"kubernetes.io/projected/d63938dd-7154-4aac-9071-59a7803032c9-kube-api-access-x8xxk\") pod \"machine-config-controller-84d6567774-ps582\" (UID: \"d63938dd-7154-4aac-9071-59a7803032c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.097151 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xff6\" (UniqueName: \"kubernetes.io/projected/090539a8-ec0a-4406-a781-2d45333291a1-kube-api-access-4xff6\") pod \"machine-config-operator-74547568cd-xxpwx\" (UID: \"090539a8-ec0a-4406-a781-2d45333291a1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:53 crc kubenswrapper[4824]: W1124 13:16:53.106199 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod774e51b9_8e46_458f_82d2_e70027a8a0fc.slice/crio-fca4985225be751a6eb8f1aa0e1af6488df6bdb0afa067b4fac21b0a4e967e8c WatchSource:0}: Error finding container fca4985225be751a6eb8f1aa0e1af6488df6bdb0afa067b4fac21b0a4e967e8c: Status 404 returned error can't find the container with id fca4985225be751a6eb8f1aa0e1af6488df6bdb0afa067b4fac21b0a4e967e8c Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.112941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.117462 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrrjn\" (UniqueName: \"kubernetes.io/projected/adc0b256-e903-4e81-8d37-8f44d27234d1-kube-api-access-lrrjn\") pod \"etcd-operator-b45778765-j9k4z\" (UID: \"adc0b256-e903-4e81-8d37-8f44d27234d1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.123899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.167999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzvxc\" (UniqueName: \"kubernetes.io/projected/2beaa442-848d-433a-841d-a461328b0376-kube-api-access-jzvxc\") pod \"collect-profiles-29399835-qbzcz\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba6e7ecf-ad36-4dd9-bfce-e5c240c07054-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nh8dl\" (UID: \"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168218 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/755c7bea-194d-4956-b24b-e2e65f031f38-metrics-tls\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168276 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/755c7bea-194d-4956-b24b-e2e65f031f38-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168300 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72e37f91-f146-4005-a77f-dc7fba802ee1-config\") pod \"kube-controller-manager-operator-78b949d7b-pbpx4\" (UID: \"72e37f91-f146-4005-a77f-dc7fba802ee1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168327 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-certificates\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72e37f91-f146-4005-a77f-dc7fba802ee1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pbpx4\" (UID: \"72e37f91-f146-4005-a77f-dc7fba802ee1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168357 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b36ecad7-891d-4814-8a54-6a3c196c0525-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168374 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3bce575-3b95-4fb4-aec2-8a3461b330d5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s96pq\" (UID: \"a3bce575-3b95-4fb4-aec2-8a3461b330d5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168392 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3f3d1884-734b-4dd3-a795-356a8e561d7b-stats-auth\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168417 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f3d1884-734b-4dd3-a795-356a8e561d7b-metrics-certs\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168443 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2-profile-collector-cert\") pod \"catalog-operator-68c6474976-sh749\" (UID: \"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168461 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2cwg\" (UniqueName: \"kubernetes.io/projected/090e8bea-a7c9-49cd-89dc-c2b0643e2bf0-kube-api-access-h2cwg\") pod \"kube-storage-version-migrator-operator-b67b599dd-rdrk8\" (UID: \"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhq8h\" (UniqueName: \"kubernetes.io/projected/3f3d1884-734b-4dd3-a795-356a8e561d7b-kube-api-access-zhq8h\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/090e8bea-a7c9-49cd-89dc-c2b0643e2bf0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rdrk8\" (UID: \"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk2c6\" (UniqueName: \"kubernetes.io/projected/fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2-kube-api-access-lk2c6\") pod \"catalog-operator-68c6474976-sh749\" (UID: \"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168565 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3bce575-3b95-4fb4-aec2-8a3461b330d5-srv-cert\") pod \"olm-operator-6b444d44fb-s96pq\" (UID: \"a3bce575-3b95-4fb4-aec2-8a3461b330d5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168581 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72e37f91-f146-4005-a77f-dc7fba802ee1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pbpx4\" (UID: \"72e37f91-f146-4005-a77f-dc7fba802ee1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168615 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3f3d1884-734b-4dd3-a795-356a8e561d7b-default-certificate\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168632 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba6e7ecf-ad36-4dd9-bfce-e5c240c07054-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nh8dl\" (UID: \"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168648 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt46b\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-kube-api-access-pt46b\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168663 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beaa442-848d-433a-841d-a461328b0376-config-volume\") pod \"collect-profiles-29399835-qbzcz\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168710 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6txs\" (UniqueName: \"kubernetes.io/projected/a3bce575-3b95-4fb4-aec2-8a3461b330d5-kube-api-access-t6txs\") pod \"olm-operator-6b444d44fb-s96pq\" (UID: \"a3bce575-3b95-4fb4-aec2-8a3461b330d5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpnvc\" (UniqueName: \"kubernetes.io/projected/755c7bea-194d-4956-b24b-e2e65f031f38-kube-api-access-bpnvc\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168774 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/755c7bea-194d-4956-b24b-e2e65f031f38-trusted-ca\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beaa442-848d-433a-841d-a461328b0376-secret-volume\") pod \"collect-profiles-29399835-qbzcz\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168833 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-tls\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168864 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-trusted-ca\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168883 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba6e7ecf-ad36-4dd9-bfce-e5c240c07054-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nh8dl\" (UID: \"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2-srv-cert\") pod \"catalog-operator-68c6474976-sh749\" (UID: \"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168951 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f3d1884-734b-4dd3-a795-356a8e561d7b-service-ca-bundle\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.168986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/090e8bea-a7c9-49cd-89dc-c2b0643e2bf0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rdrk8\" (UID: \"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.169005 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b36ecad7-891d-4814-8a54-6a3c196c0525-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.169034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-bound-sa-token\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.169103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.176103 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:53.676090829 +0000 UTC m=+155.315630139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276578 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48817d81-2ec0-4b16-b47c-b19e29b17b62-serving-cert\") pod \"service-ca-operator-777779d784-ftftq\" (UID: \"48817d81-2ec0-4b16-b47c-b19e29b17b62\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276601 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/755c7bea-194d-4956-b24b-e2e65f031f38-trusted-ca\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276628 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-csi-data-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beaa442-848d-433a-841d-a461328b0376-secret-volume\") pod \"collect-profiles-29399835-qbzcz\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276680 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-tls\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-trusted-ca\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276713 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba6e7ecf-ad36-4dd9-bfce-e5c240c07054-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nh8dl\" (UID: \"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276732 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2-srv-cert\") pod \"catalog-operator-68c6474976-sh749\" (UID: \"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9fb236a1-5efc-4aa7-9c19-692640304826-node-bootstrap-token\") pod \"machine-config-server-7dczj\" (UID: \"9fb236a1-5efc-4aa7-9c19-692640304826\") " pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48817d81-2ec0-4b16-b47c-b19e29b17b62-config\") pod \"service-ca-operator-777779d784-ftftq\" (UID: \"48817d81-2ec0-4b16-b47c-b19e29b17b62\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276780 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szkrh\" (UniqueName: \"kubernetes.io/projected/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-kube-api-access-szkrh\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f3d1884-734b-4dd3-a795-356a8e561d7b-service-ca-bundle\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/090e8bea-a7c9-49cd-89dc-c2b0643e2bf0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rdrk8\" (UID: \"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2g2n\" (UniqueName: \"kubernetes.io/projected/7c508dba-0fa1-4f58-a939-061ff854c05c-kube-api-access-h2g2n\") pod \"multus-admission-controller-857f4d67dd-mzl5h\" (UID: \"7c508dba-0fa1-4f58-a939-061ff854c05c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b36ecad7-891d-4814-8a54-6a3c196c0525-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-bound-sa-token\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d01552e-05e5-4c7a-a32f-a1fcd31d748f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ml8pn\" (UID: \"3d01552e-05e5-4c7a-a32f-a1fcd31d748f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276937 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plhqj\" (UniqueName: \"kubernetes.io/projected/91b65029-b851-4317-a1fc-26b30a5c8734-kube-api-access-plhqj\") pod \"control-plane-machine-set-operator-78cbb6b69f-8h66q\" (UID: \"91b65029-b851-4317-a1fc-26b30a5c8734\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c508dba-0fa1-4f58-a939-061ff854c05c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mzl5h\" (UID: \"7c508dba-0fa1-4f58-a939-061ff854c05c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.276994 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4fpg6\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277037 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3af944bb-d753-4fbe-a96c-20de1882117d-signing-key\") pod \"service-ca-9c57cc56f-zq9w9\" (UID: \"3af944bb-d753-4fbe-a96c-20de1882117d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzvxc\" (UniqueName: \"kubernetes.io/projected/2beaa442-848d-433a-841d-a461328b0376-kube-api-access-jzvxc\") pod \"collect-profiles-29399835-qbzcz\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-apiservice-cert\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba6e7ecf-ad36-4dd9-bfce-e5c240c07054-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nh8dl\" (UID: \"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/755c7bea-194d-4956-b24b-e2e65f031f38-metrics-tls\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277195 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f457f2da-124f-43cd-ba58-1deddaff79d7-config-volume\") pod \"dns-default-vmp98\" (UID: \"f457f2da-124f-43cd-ba58-1deddaff79d7\") " pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/755c7bea-194d-4956-b24b-e2e65f031f38-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277238 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72e37f91-f146-4005-a77f-dc7fba802ee1-config\") pod \"kube-controller-manager-operator-78b949d7b-pbpx4\" (UID: \"72e37f91-f146-4005-a77f-dc7fba802ee1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/002e2f07-c190-4749-aadf-3424ad02d9f0-cert\") pod \"ingress-canary-fnqzp\" (UID: \"002e2f07-c190-4749-aadf-3424ad02d9f0\") " pod="openshift-ingress-canary/ingress-canary-fnqzp" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn826\" (UniqueName: \"kubernetes.io/projected/48817d81-2ec0-4b16-b47c-b19e29b17b62-kube-api-access-zn826\") pod \"service-ca-operator-777779d784-ftftq\" (UID: \"48817d81-2ec0-4b16-b47c-b19e29b17b62\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277317 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/91b65029-b851-4317-a1fc-26b30a5c8734-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8h66q\" (UID: \"91b65029-b851-4317-a1fc-26b30a5c8734\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277363 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhlzr\" (UniqueName: \"kubernetes.io/projected/ecdd339c-197e-4525-8914-94be40499501-kube-api-access-xhlzr\") pod \"migrator-59844c95c7-rmdvn\" (UID: \"ecdd339c-197e-4525-8914-94be40499501\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bsw6\" (UniqueName: \"kubernetes.io/projected/3d01552e-05e5-4c7a-a32f-a1fcd31d748f-kube-api-access-5bsw6\") pod \"package-server-manager-789f6589d5-ml8pn\" (UID: \"3d01552e-05e5-4c7a-a32f-a1fcd31d748f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-certificates\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277441 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72e37f91-f146-4005-a77f-dc7fba802ee1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pbpx4\" (UID: \"72e37f91-f146-4005-a77f-dc7fba802ee1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-tmpfs\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277487 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9fb236a1-5efc-4aa7-9c19-692640304826-certs\") pod \"machine-config-server-7dczj\" (UID: \"9fb236a1-5efc-4aa7-9c19-692640304826\") " pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brbhq\" (UniqueName: \"kubernetes.io/projected/cd954f01-b73c-4c86-bebf-59fa962e70fe-kube-api-access-brbhq\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b36ecad7-891d-4814-8a54-6a3c196c0525-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f457f2da-124f-43cd-ba58-1deddaff79d7-metrics-tls\") pod \"dns-default-vmp98\" (UID: \"f457f2da-124f-43cd-ba58-1deddaff79d7\") " pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277584 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3bce575-3b95-4fb4-aec2-8a3461b330d5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s96pq\" (UID: \"a3bce575-3b95-4fb4-aec2-8a3461b330d5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277605 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-webhook-cert\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.277641 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3f3d1884-734b-4dd3-a795-356a8e561d7b-stats-auth\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.280385 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/755c7bea-194d-4956-b24b-e2e65f031f38-trusted-ca\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.280600 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlcr6\" (UniqueName: \"kubernetes.io/projected/002e2f07-c190-4749-aadf-3424ad02d9f0-kube-api-access-rlcr6\") pod \"ingress-canary-fnqzp\" (UID: \"002e2f07-c190-4749-aadf-3424ad02d9f0\") " pod="openshift-ingress-canary/ingress-canary-fnqzp" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.280865 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f3d1884-734b-4dd3-a795-356a8e561d7b-metrics-certs\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.281009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-registration-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.281261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3af944bb-d753-4fbe-a96c-20de1882117d-signing-cabundle\") pod \"service-ca-9c57cc56f-zq9w9\" (UID: \"3af944bb-d753-4fbe-a96c-20de1882117d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.282321 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2-profile-collector-cert\") pod \"catalog-operator-68c6474976-sh749\" (UID: \"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.282540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l6dh\" (UniqueName: \"kubernetes.io/projected/9fb236a1-5efc-4aa7-9c19-692640304826-kube-api-access-6l6dh\") pod \"machine-config-server-7dczj\" (UID: \"9fb236a1-5efc-4aa7-9c19-692640304826\") " pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.282575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rms8p\" (UniqueName: \"kubernetes.io/projected/3af944bb-d753-4fbe-a96c-20de1882117d-kube-api-access-rms8p\") pod \"service-ca-9c57cc56f-zq9w9\" (UID: \"3af944bb-d753-4fbe-a96c-20de1882117d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.282609 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2cwg\" (UniqueName: \"kubernetes.io/projected/090e8bea-a7c9-49cd-89dc-c2b0643e2bf0-kube-api-access-h2cwg\") pod \"kube-storage-version-migrator-operator-b67b599dd-rdrk8\" (UID: \"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.282639 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhq8h\" (UniqueName: \"kubernetes.io/projected/3f3d1884-734b-4dd3-a795-356a8e561d7b-kube-api-access-zhq8h\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.282667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/090e8bea-a7c9-49cd-89dc-c2b0643e2bf0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rdrk8\" (UID: \"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.282690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4fpg6\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.285378 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba6e7ecf-ad36-4dd9-bfce-e5c240c07054-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nh8dl\" (UID: \"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.285888 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/090e8bea-a7c9-49cd-89dc-c2b0643e2bf0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rdrk8\" (UID: \"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.286409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f3d1884-734b-4dd3-a795-356a8e561d7b-service-ca-bundle\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.288046 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-trusted-ca\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.289415 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72e37f91-f146-4005-a77f-dc7fba802ee1-config\") pod \"kube-controller-manager-operator-78b949d7b-pbpx4\" (UID: \"72e37f91-f146-4005-a77f-dc7fba802ee1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.290616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-certificates\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.290757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b36ecad7-891d-4814-8a54-6a3c196c0525-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.302516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk2c6\" (UniqueName: \"kubernetes.io/projected/fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2-kube-api-access-lk2c6\") pod \"catalog-operator-68c6474976-sh749\" (UID: \"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.302586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3bce575-3b95-4fb4-aec2-8a3461b330d5-srv-cert\") pod \"olm-operator-6b444d44fb-s96pq\" (UID: \"a3bce575-3b95-4fb4-aec2-8a3461b330d5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.302618 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3f3d1884-734b-4dd3-a795-356a8e561d7b-default-certificate\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.302743 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:53.802710654 +0000 UTC m=+155.442249964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.302834 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72e37f91-f146-4005-a77f-dc7fba802ee1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pbpx4\" (UID: \"72e37f91-f146-4005-a77f-dc7fba802ee1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.302865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-socket-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.302893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-mountpoint-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.302921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba6e7ecf-ad36-4dd9-bfce-e5c240c07054-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nh8dl\" (UID: \"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.302975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beaa442-848d-433a-841d-a461328b0376-config-volume\") pod \"collect-profiles-29399835-qbzcz\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.302999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-plugins-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.303028 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt46b\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-kube-api-access-pt46b\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.303072 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2d9n\" (UniqueName: \"kubernetes.io/projected/d41e0dcc-fb19-49fc-8aab-a1837193c058-kube-api-access-h2d9n\") pod \"marketplace-operator-79b997595-4fpg6\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.314074 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3bce575-3b95-4fb4-aec2-8a3461b330d5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s96pq\" (UID: \"a3bce575-3b95-4fb4-aec2-8a3461b330d5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.319446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beaa442-848d-433a-841d-a461328b0376-config-volume\") pod \"collect-profiles-29399835-qbzcz\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.321223 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/755c7bea-194d-4956-b24b-e2e65f031f38-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.321695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6txs\" (UniqueName: \"kubernetes.io/projected/a3bce575-3b95-4fb4-aec2-8a3461b330d5-kube-api-access-t6txs\") pod \"olm-operator-6b444d44fb-s96pq\" (UID: \"a3bce575-3b95-4fb4-aec2-8a3461b330d5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.321720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wjml\" (UniqueName: \"kubernetes.io/projected/f457f2da-124f-43cd-ba58-1deddaff79d7-kube-api-access-7wjml\") pod \"dns-default-vmp98\" (UID: \"f457f2da-124f-43cd-ba58-1deddaff79d7\") " pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.321861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpnvc\" (UniqueName: \"kubernetes.io/projected/755c7bea-194d-4956-b24b-e2e65f031f38-kube-api-access-bpnvc\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.325014 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3bce575-3b95-4fb4-aec2-8a3461b330d5-srv-cert\") pod \"olm-operator-6b444d44fb-s96pq\" (UID: \"a3bce575-3b95-4fb4-aec2-8a3461b330d5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.326610 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3f3d1884-734b-4dd3-a795-356a8e561d7b-stats-auth\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.340714 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzvxc\" (UniqueName: \"kubernetes.io/projected/2beaa442-848d-433a-841d-a461328b0376-kube-api-access-jzvxc\") pod \"collect-profiles-29399835-qbzcz\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.344663 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" event={"ID":"831ebed0-2a1f-490a-b315-27f535e22fe0","Type":"ContainerStarted","Data":"657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.345581 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.348407 4824 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-f9ngl container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.348454 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" podUID="831ebed0-2a1f-490a-b315-27f535e22fe0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.351749 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/090e8bea-a7c9-49cd-89dc-c2b0643e2bf0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rdrk8\" (UID: \"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.351748 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/755c7bea-194d-4956-b24b-e2e65f031f38-metrics-tls\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.352252 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2-srv-cert\") pod \"catalog-operator-68c6474976-sh749\" (UID: \"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.352253 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-tls\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.352506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72e37f91-f146-4005-a77f-dc7fba802ee1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pbpx4\" (UID: \"72e37f91-f146-4005-a77f-dc7fba802ee1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.352723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2-profile-collector-cert\") pod \"catalog-operator-68c6474976-sh749\" (UID: \"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.354409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-bound-sa-token\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.357108 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba6e7ecf-ad36-4dd9-bfce-e5c240c07054-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nh8dl\" (UID: \"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.365710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3f3d1884-734b-4dd3-a795-356a8e561d7b-default-certificate\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.365983 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.366451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b36ecad7-891d-4814-8a54-6a3c196c0525-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.366558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beaa442-848d-433a-841d-a461328b0376-secret-volume\") pod \"collect-profiles-29399835-qbzcz\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.372241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f3d1884-734b-4dd3-a795-356a8e561d7b-metrics-certs\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.372345 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba6e7ecf-ad36-4dd9-bfce-e5c240c07054-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nh8dl\" (UID: \"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.376202 4824 generic.go:334] "Generic (PLEG): container finished" podID="5efdc9d2-2541-4550-9f68-485fbc54473f" containerID="6c0f1932cf33aef3d4bb08bef2b179b1a669b34b9bcfd76982d6997ca87e880f" exitCode=0 Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.376296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" event={"ID":"5efdc9d2-2541-4550-9f68-485fbc54473f","Type":"ContainerDied","Data":"6c0f1932cf33aef3d4bb08bef2b179b1a669b34b9bcfd76982d6997ca87e880f"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.376321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" event={"ID":"5efdc9d2-2541-4550-9f68-485fbc54473f","Type":"ContainerStarted","Data":"08819c7c6ec19f66c63bdbf27084333da868237bc6d07b15ea35e56db5472bc5"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.383425 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.386477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" event={"ID":"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586","Type":"ContainerStarted","Data":"e479687978a1bc0efcc2d5a6ad4eb99e3edc2f216d70784febbafa39aa216983"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.386566 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" event={"ID":"be5b6bdb-7ccd-4401-a46e-4c67fcd1c586","Type":"ContainerStarted","Data":"2b5b1ff10c07fe2de9fa57aebbbf3af55d8aea5c5927ddef2b9070a4c4071c37"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.407210 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2cwg\" (UniqueName: \"kubernetes.io/projected/090e8bea-a7c9-49cd-89dc-c2b0643e2bf0-kube-api-access-h2cwg\") pod \"kube-storage-version-migrator-operator-b67b599dd-rdrk8\" (UID: \"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.407624 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.417548 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhq8h\" (UniqueName: \"kubernetes.io/projected/3f3d1884-734b-4dd3-a795-356a8e561d7b-kube-api-access-zhq8h\") pod \"router-default-5444994796-7lwbq\" (UID: \"3f3d1884-734b-4dd3-a795-356a8e561d7b\") " pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn826\" (UniqueName: \"kubernetes.io/projected/48817d81-2ec0-4b16-b47c-b19e29b17b62-kube-api-access-zn826\") pod \"service-ca-operator-777779d784-ftftq\" (UID: \"48817d81-2ec0-4b16-b47c-b19e29b17b62\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/91b65029-b851-4317-a1fc-26b30a5c8734-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8h66q\" (UID: \"91b65029-b851-4317-a1fc-26b30a5c8734\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhlzr\" (UniqueName: \"kubernetes.io/projected/ecdd339c-197e-4525-8914-94be40499501-kube-api-access-xhlzr\") pod \"migrator-59844c95c7-rmdvn\" (UID: \"ecdd339c-197e-4525-8914-94be40499501\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bsw6\" (UniqueName: \"kubernetes.io/projected/3d01552e-05e5-4c7a-a32f-a1fcd31d748f-kube-api-access-5bsw6\") pod \"package-server-manager-789f6589d5-ml8pn\" (UID: \"3d01552e-05e5-4c7a-a32f-a1fcd31d748f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423467 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9fb236a1-5efc-4aa7-9c19-692640304826-certs\") pod \"machine-config-server-7dczj\" (UID: \"9fb236a1-5efc-4aa7-9c19-692640304826\") " pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423486 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-tmpfs\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423506 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brbhq\" (UniqueName: \"kubernetes.io/projected/cd954f01-b73c-4c86-bebf-59fa962e70fe-kube-api-access-brbhq\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f457f2da-124f-43cd-ba58-1deddaff79d7-metrics-tls\") pod \"dns-default-vmp98\" (UID: \"f457f2da-124f-43cd-ba58-1deddaff79d7\") " pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-webhook-cert\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlcr6\" (UniqueName: \"kubernetes.io/projected/002e2f07-c190-4749-aadf-3424ad02d9f0-kube-api-access-rlcr6\") pod \"ingress-canary-fnqzp\" (UID: \"002e2f07-c190-4749-aadf-3424ad02d9f0\") " pod="openshift-ingress-canary/ingress-canary-fnqzp" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423599 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-registration-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l6dh\" (UniqueName: \"kubernetes.io/projected/9fb236a1-5efc-4aa7-9c19-692640304826-kube-api-access-6l6dh\") pod \"machine-config-server-7dczj\" (UID: \"9fb236a1-5efc-4aa7-9c19-692640304826\") " pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3af944bb-d753-4fbe-a96c-20de1882117d-signing-cabundle\") pod \"service-ca-9c57cc56f-zq9w9\" (UID: \"3af944bb-d753-4fbe-a96c-20de1882117d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rms8p\" (UniqueName: \"kubernetes.io/projected/3af944bb-d753-4fbe-a96c-20de1882117d-kube-api-access-rms8p\") pod \"service-ca-9c57cc56f-zq9w9\" (UID: \"3af944bb-d753-4fbe-a96c-20de1882117d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423757 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4fpg6\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423796 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-socket-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-mountpoint-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423901 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-plugins-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2d9n\" (UniqueName: \"kubernetes.io/projected/d41e0dcc-fb19-49fc-8aab-a1837193c058-kube-api-access-h2d9n\") pod \"marketplace-operator-79b997595-4fpg6\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423947 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wjml\" (UniqueName: \"kubernetes.io/projected/f457f2da-124f-43cd-ba58-1deddaff79d7-kube-api-access-7wjml\") pod \"dns-default-vmp98\" (UID: \"f457f2da-124f-43cd-ba58-1deddaff79d7\") " pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.423989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48817d81-2ec0-4b16-b47c-b19e29b17b62-serving-cert\") pod \"service-ca-operator-777779d784-ftftq\" (UID: \"48817d81-2ec0-4b16-b47c-b19e29b17b62\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424009 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-csi-data-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424047 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9fb236a1-5efc-4aa7-9c19-692640304826-node-bootstrap-token\") pod \"machine-config-server-7dczj\" (UID: \"9fb236a1-5efc-4aa7-9c19-692640304826\") " pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424065 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48817d81-2ec0-4b16-b47c-b19e29b17b62-config\") pod \"service-ca-operator-777779d784-ftftq\" (UID: \"48817d81-2ec0-4b16-b47c-b19e29b17b62\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szkrh\" (UniqueName: \"kubernetes.io/projected/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-kube-api-access-szkrh\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424126 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2g2n\" (UniqueName: \"kubernetes.io/projected/7c508dba-0fa1-4f58-a939-061ff854c05c-kube-api-access-h2g2n\") pod \"multus-admission-controller-857f4d67dd-mzl5h\" (UID: \"7c508dba-0fa1-4f58-a939-061ff854c05c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d01552e-05e5-4c7a-a32f-a1fcd31d748f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ml8pn\" (UID: \"3d01552e-05e5-4c7a-a32f-a1fcd31d748f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plhqj\" (UniqueName: \"kubernetes.io/projected/91b65029-b851-4317-a1fc-26b30a5c8734-kube-api-access-plhqj\") pod \"control-plane-machine-set-operator-78cbb6b69f-8h66q\" (UID: \"91b65029-b851-4317-a1fc-26b30a5c8734\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c508dba-0fa1-4f58-a939-061ff854c05c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mzl5h\" (UID: \"7c508dba-0fa1-4f58-a939-061ff854c05c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4fpg6\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424272 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424291 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3af944bb-d753-4fbe-a96c-20de1882117d-signing-key\") pod \"service-ca-9c57cc56f-zq9w9\" (UID: \"3af944bb-d753-4fbe-a96c-20de1882117d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-apiservice-cert\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424342 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f457f2da-124f-43cd-ba58-1deddaff79d7-config-volume\") pod \"dns-default-vmp98\" (UID: \"f457f2da-124f-43cd-ba58-1deddaff79d7\") " pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/002e2f07-c190-4749-aadf-3424ad02d9f0-cert\") pod \"ingress-canary-fnqzp\" (UID: \"002e2f07-c190-4749-aadf-3424ad02d9f0\") " pod="openshift-ingress-canary/ingress-canary-fnqzp" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.424513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-mountpoint-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.427221 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-plugins-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.428147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-csi-data-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.430201 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3af944bb-d753-4fbe-a96c-20de1882117d-signing-cabundle\") pod \"service-ca-9c57cc56f-zq9w9\" (UID: \"3af944bb-d753-4fbe-a96c-20de1882117d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.430699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-registration-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.431068 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4fpg6\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.431592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/cd954f01-b73c-4c86-bebf-59fa962e70fe-socket-dir\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.434436 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48817d81-2ec0-4b16-b47c-b19e29b17b62-config\") pod \"service-ca-operator-777779d784-ftftq\" (UID: \"48817d81-2ec0-4b16-b47c-b19e29b17b62\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.435889 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:53.935864661 +0000 UTC m=+155.575403971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.440470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-tmpfs\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.441024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d01552e-05e5-4c7a-a32f-a1fcd31d748f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ml8pn\" (UID: \"3d01552e-05e5-4c7a-a32f-a1fcd31d748f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.442883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48817d81-2ec0-4b16-b47c-b19e29b17b62-serving-cert\") pod \"service-ca-operator-777779d784-ftftq\" (UID: \"48817d81-2ec0-4b16-b47c-b19e29b17b62\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.443976 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f457f2da-124f-43cd-ba58-1deddaff79d7-config-volume\") pod \"dns-default-vmp98\" (UID: \"f457f2da-124f-43cd-ba58-1deddaff79d7\") " pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.446656 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.447593 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/91b65029-b851-4317-a1fc-26b30a5c8734-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8h66q\" (UID: \"91b65029-b851-4317-a1fc-26b30a5c8734\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.448880 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" event={"ID":"8a59be8e-4194-4e1c-89bd-3136366ed13b","Type":"ContainerStarted","Data":"6a92bdd70d35e73d690c6f1739214c42682914b099d81f45f88e6dd8bd6088c9"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.448932 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" event={"ID":"8a59be8e-4194-4e1c-89bd-3136366ed13b","Type":"ContainerStarted","Data":"22127108569b48e8fcabbce3acb85b4e612745d968f73a532d40a7b4b7874abe"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.448947 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-webhook-cert\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.460465 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dw5hv"] Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.475375 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72e37f91-f146-4005-a77f-dc7fba802ee1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pbpx4\" (UID: \"72e37f91-f146-4005-a77f-dc7fba802ee1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.487205 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-apiservice-cert\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.487791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3af944bb-d753-4fbe-a96c-20de1882117d-signing-key\") pod \"service-ca-9c57cc56f-zq9w9\" (UID: \"3af944bb-d753-4fbe-a96c-20de1882117d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.488770 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk2c6\" (UniqueName: \"kubernetes.io/projected/fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2-kube-api-access-lk2c6\") pod \"catalog-operator-68c6474976-sh749\" (UID: \"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.489757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f457f2da-124f-43cd-ba58-1deddaff79d7-metrics-tls\") pod \"dns-default-vmp98\" (UID: \"f457f2da-124f-43cd-ba58-1deddaff79d7\") " pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.491400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt46b\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-kube-api-access-pt46b\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.499431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9fb236a1-5efc-4aa7-9c19-692640304826-certs\") pod \"machine-config-server-7dczj\" (UID: \"9fb236a1-5efc-4aa7-9c19-692640304826\") " pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.500063 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6txs\" (UniqueName: \"kubernetes.io/projected/a3bce575-3b95-4fb4-aec2-8a3461b330d5-kube-api-access-t6txs\") pod \"olm-operator-6b444d44fb-s96pq\" (UID: \"a3bce575-3b95-4fb4-aec2-8a3461b330d5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.501791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c508dba-0fa1-4f58-a939-061ff854c05c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mzl5h\" (UID: \"7c508dba-0fa1-4f58-a939-061ff854c05c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.508023 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/002e2f07-c190-4749-aadf-3424ad02d9f0-cert\") pod \"ingress-canary-fnqzp\" (UID: \"002e2f07-c190-4749-aadf-3424ad02d9f0\") " pod="openshift-ingress-canary/ingress-canary-fnqzp" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.508681 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9fb236a1-5efc-4aa7-9c19-692640304826-node-bootstrap-token\") pod \"machine-config-server-7dczj\" (UID: \"9fb236a1-5efc-4aa7-9c19-692640304826\") " pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.523587 4824 generic.go:334] "Generic (PLEG): container finished" podID="64d54af6-d194-48c1-9630-58d50c690eb3" containerID="2cd7dc54763f04d3c9ac871ca4bfa33c53383facae812cc703296417c16bd9e1" exitCode=0 Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.523656 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" event={"ID":"64d54af6-d194-48c1-9630-58d50c690eb3","Type":"ContainerDied","Data":"2cd7dc54763f04d3c9ac871ca4bfa33c53383facae812cc703296417c16bd9e1"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.523682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" event={"ID":"64d54af6-d194-48c1-9630-58d50c690eb3","Type":"ContainerStarted","Data":"b52fb883024de3e669d78a0ca8d585a59fc7e394ec06e9f5a5edafb4f2593a86"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.525050 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.528343 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpnvc\" (UniqueName: \"kubernetes.io/projected/755c7bea-194d-4956-b24b-e2e65f031f38-kube-api-access-bpnvc\") pod \"ingress-operator-5b745b69d9-kpqbs\" (UID: \"755c7bea-194d-4956-b24b-e2e65f031f38\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.529408 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.029388349 +0000 UTC m=+155.668927659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.530581 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4fpg6\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.534342 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.534675 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.03466278 +0000 UTC m=+155.674202090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.565739 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7"] Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.569610 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn826\" (UniqueName: \"kubernetes.io/projected/48817d81-2ec0-4b16-b47c-b19e29b17b62-kube-api-access-zn826\") pod \"service-ca-operator-777779d784-ftftq\" (UID: \"48817d81-2ec0-4b16-b47c-b19e29b17b62\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.588142 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" event={"ID":"f4875b41-d0f2-48c9-a451-839cd764a9cb","Type":"ContainerStarted","Data":"0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.588189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" event={"ID":"f4875b41-d0f2-48c9-a451-839cd764a9cb","Type":"ContainerStarted","Data":"8db400acca7af97b15189fd31ec4f87e90b9d9fa4c4015b9f5beea4647838487"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.590356 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.593106 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2g2n\" (UniqueName: \"kubernetes.io/projected/7c508dba-0fa1-4f58-a939-061ff854c05c-kube-api-access-h2g2n\") pod \"multus-admission-controller-857f4d67dd-mzl5h\" (UID: \"7c508dba-0fa1-4f58-a939-061ff854c05c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.608822 4824 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-5p99t container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.608884 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" podUID="f4875b41-d0f2-48c9-a451-839cd764a9cb" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.612202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" event={"ID":"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1","Type":"ContainerStarted","Data":"46a5f973dd55d4779b7972ff1ae4cf6849ddb0f026d69cb7708d93d11e2370ec"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.612256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" event={"ID":"e7abfe4e-76f0-4a5a-9dcd-50a80dd8d3e1","Type":"ContainerStarted","Data":"cfd592a9038b08454a0c4effa9d4ac8695f67fa2e23a9d24ec9eb6e1c7493593"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.613655 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2d9n\" (UniqueName: \"kubernetes.io/projected/d41e0dcc-fb19-49fc-8aab-a1837193c058-kube-api-access-h2d9n\") pod \"marketplace-operator-79b997595-4fpg6\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.616946 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" event={"ID":"fcf8e28b-55ce-4fd8-891b-50e2381d29a4","Type":"ContainerStarted","Data":"2706276d55f1b051ca6427f4edaa464b7514f17edf0ed9902306df0785876e87"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.616991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" event={"ID":"fcf8e28b-55ce-4fd8-891b-50e2381d29a4","Type":"ContainerStarted","Data":"3377ec2844d0cb024f617e21d06910bc2858f7b9dd1a82440686b56d3a2a0281"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.618161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wjml\" (UniqueName: \"kubernetes.io/projected/f457f2da-124f-43cd-ba58-1deddaff79d7-kube-api-access-7wjml\") pod \"dns-default-vmp98\" (UID: \"f457f2da-124f-43cd-ba58-1deddaff79d7\") " pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.638126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.638488 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.138458364 +0000 UTC m=+155.777997674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.638657 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.639539 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.139527391 +0000 UTC m=+155.779066701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.639998 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9"] Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.643919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plhqj\" (UniqueName: \"kubernetes.io/projected/91b65029-b851-4317-a1fc-26b30a5c8734-kube-api-access-plhqj\") pod \"control-plane-machine-set-operator-78cbb6b69f-8h66q\" (UID: \"91b65029-b851-4317-a1fc-26b30a5c8734\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.644306 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" event={"ID":"93a94b2f-1727-44d6-8438-c1492485cf06","Type":"ContainerStarted","Data":"444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.644381 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" event={"ID":"93a94b2f-1727-44d6-8438-c1492485cf06","Type":"ContainerStarted","Data":"770475f891428760a1ec46278bb77e6dc845deb090e432c79f48d86402599dab"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.646170 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.652117 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.652482 4824 generic.go:334] "Generic (PLEG): container finished" podID="60b8f545-2446-4422-8838-ac6b54ea119f" containerID="40b30a71fa9a586ee07fd374a12605d4b95c151a97f8ee5f9de41f1a01d87cef" exitCode=0 Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.652504 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" event={"ID":"60b8f545-2446-4422-8838-ac6b54ea119f","Type":"ContainerDied","Data":"40b30a71fa9a586ee07fd374a12605d4b95c151a97f8ee5f9de41f1a01d87cef"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.655043 4824 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lm6d6 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.655080 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" podUID="93a94b2f-1727-44d6-8438-c1492485cf06" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.660611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" event={"ID":"774e51b9-8e46-458f-82d2-e70027a8a0fc","Type":"ContainerStarted","Data":"fca4985225be751a6eb8f1aa0e1af6488df6bdb0afa067b4fac21b0a4e967e8c"} Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.664750 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szkrh\" (UniqueName: \"kubernetes.io/projected/4cf5d5ca-477f-4750-8cd1-78a1db27f9d6-kube-api-access-szkrh\") pod \"packageserver-d55dfcdfc-rl9nn\" (UID: \"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.683078 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlcr6\" (UniqueName: \"kubernetes.io/projected/002e2f07-c190-4749-aadf-3424ad02d9f0-kube-api-access-rlcr6\") pod \"ingress-canary-fnqzp\" (UID: \"002e2f07-c190-4749-aadf-3424ad02d9f0\") " pod="openshift-ingress-canary/ingress-canary-fnqzp" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.690218 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.696646 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:53 crc kubenswrapper[4824]: W1124 13:16:53.697017 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01ec77ac_5384_4457_8d8a_9d14135c2b82.slice/crio-c35222676e11b7f282071dc4da29002b29dc816b51625ec97cdc47bca819498a WatchSource:0}: Error finding container c35222676e11b7f282071dc4da29002b29dc816b51625ec97cdc47bca819498a: Status 404 returned error can't find the container with id c35222676e11b7f282071dc4da29002b29dc816b51625ec97cdc47bca819498a Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.719459 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rms8p\" (UniqueName: \"kubernetes.io/projected/3af944bb-d753-4fbe-a96c-20de1882117d-kube-api-access-rms8p\") pod \"service-ca-9c57cc56f-zq9w9\" (UID: \"3af944bb-d753-4fbe-a96c-20de1882117d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.719761 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.736177 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.737255 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.737509 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l6dh\" (UniqueName: \"kubernetes.io/projected/9fb236a1-5efc-4aa7-9c19-692640304826-kube-api-access-6l6dh\") pod \"machine-config-server-7dczj\" (UID: \"9fb236a1-5efc-4aa7-9c19-692640304826\") " pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.739400 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.740781 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.240762751 +0000 UTC m=+155.880302061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.754512 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.761420 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vwk9h"] Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.765081 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.770637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brbhq\" (UniqueName: \"kubernetes.io/projected/cd954f01-b73c-4c86-bebf-59fa962e70fe-kube-api-access-brbhq\") pod \"csi-hostpathplugin-ksxcs\" (UID: \"cd954f01-b73c-4c86-bebf-59fa962e70fe\") " pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.788041 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.788996 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bsw6\" (UniqueName: \"kubernetes.io/projected/3d01552e-05e5-4c7a-a32f-a1fcd31d748f-kube-api-access-5bsw6\") pod \"package-server-manager-789f6589d5-ml8pn\" (UID: \"3d01552e-05e5-4c7a-a32f-a1fcd31d748f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.790456 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhlzr\" (UniqueName: \"kubernetes.io/projected/ecdd339c-197e-4525-8914-94be40499501-kube-api-access-xhlzr\") pod \"migrator-59844c95c7-rmdvn\" (UID: \"ecdd339c-197e-4525-8914-94be40499501\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.795300 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.819698 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.820133 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.837895 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.839675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vmp98" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.841285 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6s7gg" podStartSLOduration=128.841271843 podStartE2EDuration="2m8.841271843s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:53.839306734 +0000 UTC m=+155.478846044" watchObservedRunningTime="2025-11-24 13:16:53.841271843 +0000 UTC m=+155.480811153" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.841504 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.842734 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx"] Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.844356 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.344326669 +0000 UTC m=+155.983865979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.848195 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fnqzp" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.880858 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7dczj" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.881159 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.896681 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p"] Nov 24 13:16:53 crc kubenswrapper[4824]: I1124 13:16:53.944350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:53 crc kubenswrapper[4824]: E1124 13:16:53.944760 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.444741759 +0000 UTC m=+156.084281069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:53 crc kubenswrapper[4824]: W1124 13:16:53.972240 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod090539a8_ec0a_4406_a781_2d45333291a1.slice/crio-ad98a7e540860e5d4a4406aace36bcf038bcc30c3396c736294ccb0a9a16a48c WatchSource:0}: Error finding container ad98a7e540860e5d4a4406aace36bcf038bcc30c3396c736294ccb0a9a16a48c: Status 404 returned error can't find the container with id ad98a7e540860e5d4a4406aace36bcf038bcc30c3396c736294ccb0a9a16a48c Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.048763 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.054040 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.55401823 +0000 UTC m=+156.193557540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.054821 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ll2hq"] Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.075108 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.128107 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j9k4z"] Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.150510 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.158741 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.658709316 +0000 UTC m=+156.298248626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.173885 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ps582"] Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.176008 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.176598 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.676585363 +0000 UTC m=+156.316124673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.277467 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.278028 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.777996127 +0000 UTC m=+156.417535437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: W1124 13:16:54.368796 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76c0d257_aca8_434e_97b2_067ddc1e7f66.slice/crio-368c1773b03ecb2b9417586586be7b4aa39785ba79e860880bcada8e49fc0b04 WatchSource:0}: Error finding container 368c1773b03ecb2b9417586586be7b4aa39785ba79e860880bcada8e49fc0b04: Status 404 returned error can't find the container with id 368c1773b03ecb2b9417586586be7b4aa39785ba79e860880bcada8e49fc0b04 Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.382987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.383595 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.883579325 +0000 UTC m=+156.523118635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: W1124 13:16:54.396083 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadc0b256_e903_4e81_8d37_8f44d27234d1.slice/crio-6ea36d9cbd6b1d7f381782e26baf96f3d8f22a84c177ded3abc3986503770f92 WatchSource:0}: Error finding container 6ea36d9cbd6b1d7f381782e26baf96f3d8f22a84c177ded3abc3986503770f92: Status 404 returned error can't find the container with id 6ea36d9cbd6b1d7f381782e26baf96f3d8f22a84c177ded3abc3986503770f92 Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.476200 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-p6x6w" Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.483959 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.484643 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:54.984624771 +0000 UTC m=+156.624164081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.505969 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl"] Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.548180 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz"] Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.582529 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs"] Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.588750 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.589193 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.089177204 +0000 UTC m=+156.728716504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.594051 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5r76" podStartSLOduration=128.594033375 podStartE2EDuration="2m8.594033375s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:54.573852671 +0000 UTC m=+156.213391981" watchObservedRunningTime="2025-11-24 13:16:54.594033375 +0000 UTC m=+156.233572675" Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.594893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ftftq"] Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.689539 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.689766 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.189750127 +0000 UTC m=+156.829289437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.702401 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" event={"ID":"2cfbf6ab-805e-4776-8f3e-1823b675c1be","Type":"ContainerStarted","Data":"4bd0bcb368481c538e99b35504ab7b061d499694c204508ced530018c24735de"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.702440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" event={"ID":"2cfbf6ab-805e-4776-8f3e-1823b675c1be","Type":"ContainerStarted","Data":"ebcd6e7fd8e2c08d3bdefc60d1b04e2426f20bc2e65b179d359540f198590cbd"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.722614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" event={"ID":"adc0b256-e903-4e81-8d37-8f44d27234d1","Type":"ContainerStarted","Data":"6ea36d9cbd6b1d7f381782e26baf96f3d8f22a84c177ded3abc3986503770f92"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.731015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" event={"ID":"7e7668d3-1c2b-4c90-9940-c1898a6c5069","Type":"ContainerStarted","Data":"170eb5e17368081d7f1c774264134f025d0959b9d0279e06cfe92dbd08299fbe"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.732991 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" podStartSLOduration=128.732971737 podStartE2EDuration="2m8.732971737s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:54.71787279 +0000 UTC m=+156.357412090" watchObservedRunningTime="2025-11-24 13:16:54.732971737 +0000 UTC m=+156.372511047" Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.733255 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mzl5h"] Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.744085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vwk9h" event={"ID":"ecea50ae-16db-46d7-a9cd-84a339185ae4","Type":"ContainerStarted","Data":"6afb63632b5a6e4f7b722f97cff4503be0c196d43153f169a804e5f3e3aaf365"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.754877 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8"] Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.776722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" event={"ID":"269ad190-1f61-4290-813b-86950552ea1f","Type":"ContainerStarted","Data":"ad2b3156ad2319eecbbf60739ce47b7ed775cfab51d191f776d18f3c994a7ea6"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.776766 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" event={"ID":"269ad190-1f61-4290-813b-86950552ea1f","Type":"ContainerStarted","Data":"d69e43d0db2bfe44ee6dd6e5e33298a03d9680c127482cdc0d575bd09fd63a75"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.794131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.794480 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.294463774 +0000 UTC m=+156.934003094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.846838 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" event={"ID":"d63938dd-7154-4aac-9071-59a7803032c9","Type":"ContainerStarted","Data":"c02e67d38edbfaa3638cdec90a22a47f8b6d8d1e0148dbcc82115345ba883497"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.895769 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.897519 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.397500979 +0000 UTC m=+157.037040299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.913627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7lwbq" event={"ID":"3f3d1884-734b-4dd3-a795-356a8e561d7b","Type":"ContainerStarted","Data":"9033a7d524d4d15d7ed277d43040096cd50c7623233c8b0249f11cef96e614cf"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.962976 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" event={"ID":"090539a8-ec0a-4406-a781-2d45333291a1","Type":"ContainerStarted","Data":"ad98a7e540860e5d4a4406aace36bcf038bcc30c3396c736294ccb0a9a16a48c"} Nov 24 13:16:54 crc kubenswrapper[4824]: I1124 13:16:54.998798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:54 crc kubenswrapper[4824]: E1124 13:16:54.999092 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.499079187 +0000 UTC m=+157.138618497 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.036861 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-p6x6w" podStartSLOduration=130.036845441 podStartE2EDuration="2m10.036845441s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:55.035041466 +0000 UTC m=+156.674580796" watchObservedRunningTime="2025-11-24 13:16:55.036845441 +0000 UTC m=+156.676384751" Nov 24 13:16:55 crc kubenswrapper[4824]: W1124 13:16:55.086064 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2beaa442_848d_433a_841d_a461328b0376.slice/crio-697721b10203ed2166eea3fe24b0128a7307b85a551130b3277530249e1be722 WatchSource:0}: Error finding container 697721b10203ed2166eea3fe24b0128a7307b85a551130b3277530249e1be722: Status 404 returned error can't find the container with id 697721b10203ed2166eea3fe24b0128a7307b85a551130b3277530249e1be722 Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.096386 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" event={"ID":"01ec77ac-5384-4457-8d8a-9d14135c2b82","Type":"ContainerStarted","Data":"c35222676e11b7f282071dc4da29002b29dc816b51625ec97cdc47bca819498a"} Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.100696 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.108223 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.608192704 +0000 UTC m=+157.247732004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.108338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.108715 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.608695417 +0000 UTC m=+157.248234727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.132353 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" event={"ID":"774e51b9-8e46-458f-82d2-e70027a8a0fc","Type":"ContainerStarted","Data":"7f96be032e08cfe03cda1d1d8a69fb116a912a20fe8b7be53b4004cd876cfc06"} Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.143773 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" event={"ID":"64d54af6-d194-48c1-9630-58d50c690eb3","Type":"ContainerStarted","Data":"bfcbadd3143639de5b717d45b8f7edd09239e5c69c79313543b53a3fb7fc1407"} Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.144148 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.162756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ll2hq" event={"ID":"76c0d257-aca8-434e-97b2-067ddc1e7f66","Type":"ContainerStarted","Data":"368c1773b03ecb2b9417586586be7b4aa39785ba79e860880bcada8e49fc0b04"} Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.171870 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq"] Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.184425 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.184578 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.208974 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.209297 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.209641 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.709625659 +0000 UTC m=+157.349164969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.230266 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn"] Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.265426 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4"] Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.310333 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.312338 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.812325856 +0000 UTC m=+157.451865166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.418759 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.420389 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:55.920356886 +0000 UTC m=+157.559896196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.473865 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t9jw7" podStartSLOduration=130.473846072 podStartE2EDuration="2m10.473846072s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:55.473386691 +0000 UTC m=+157.112926001" watchObservedRunningTime="2025-11-24 13:16:55.473846072 +0000 UTC m=+157.113385382" Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.487938 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4fpg6"] Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.521354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.521720 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.021709018 +0000 UTC m=+157.661248328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: W1124 13:16:55.529643 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72e37f91_f146_4005_a77f_dc7fba802ee1.slice/crio-f603f008e803125709214ddb63e1ac77adfe3efb400fa6225d32ee0e327126f0 WatchSource:0}: Error finding container f603f008e803125709214ddb63e1ac77adfe3efb400fa6225d32ee0e327126f0: Status 404 returned error can't find the container with id f603f008e803125709214ddb63e1ac77adfe3efb400fa6225d32ee0e327126f0 Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.532354 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ksxcs"] Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.563128 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn"] Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.582797 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vd2g5" podStartSLOduration=130.582778675 podStartE2EDuration="2m10.582778675s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:55.580475447 +0000 UTC m=+157.220014757" watchObservedRunningTime="2025-11-24 13:16:55.582778675 +0000 UTC m=+157.222317985" Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.600302 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vmp98"] Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.624734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.625447 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.125431501 +0000 UTC m=+157.764970811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.730174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.730502 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.230491036 +0000 UTC m=+157.870030346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.760495 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" podStartSLOduration=130.760478565 podStartE2EDuration="2m10.760478565s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:55.676683641 +0000 UTC m=+157.316222951" watchObservedRunningTime="2025-11-24 13:16:55.760478565 +0000 UTC m=+157.400017875" Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.816084 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fnqzp"] Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.830452 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zq9w9"] Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.834027 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.834499 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.334480325 +0000 UTC m=+157.974019635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.941587 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:55 crc kubenswrapper[4824]: E1124 13:16:55.942177 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.442160465 +0000 UTC m=+158.081699775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:55 crc kubenswrapper[4824]: I1124 13:16:55.956113 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8wzb5" podStartSLOduration=130.956096733 podStartE2EDuration="2m10.956096733s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:55.867450678 +0000 UTC m=+157.506989988" watchObservedRunningTime="2025-11-24 13:16:55.956096733 +0000 UTC m=+157.595636043" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.022143 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" podStartSLOduration=131.022120623 podStartE2EDuration="2m11.022120623s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:55.958622866 +0000 UTC m=+157.598162176" watchObservedRunningTime="2025-11-24 13:16:56.022120623 +0000 UTC m=+157.661659933" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.022408 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749"] Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.046612 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:56 crc kubenswrapper[4824]: E1124 13:16:56.047124 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.547105057 +0000 UTC m=+158.186644367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.060571 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" podStartSLOduration=131.060551174 podStartE2EDuration="2m11.060551174s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:56.05841127 +0000 UTC m=+157.697950580" watchObservedRunningTime="2025-11-24 13:16:56.060551174 +0000 UTC m=+157.700090474" Nov 24 13:16:56 crc kubenswrapper[4824]: W1124 13:16:56.135943 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb7ac5a4_915f_46e6_9ba0_8e6ce76423e2.slice/crio-8e7cdd896d19e9d3374b2d843f20f5f7ae964ed50e309184c3536cf36e0169ee WatchSource:0}: Error finding container 8e7cdd896d19e9d3374b2d843f20f5f7ae964ed50e309184c3536cf36e0169ee: Status 404 returned error can't find the container with id 8e7cdd896d19e9d3374b2d843f20f5f7ae964ed50e309184c3536cf36e0169ee Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.150015 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:56 crc kubenswrapper[4824]: E1124 13:16:56.150329 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.650315347 +0000 UTC m=+158.289854657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.229085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7dczj" event={"ID":"9fb236a1-5efc-4aa7-9c19-692640304826","Type":"ContainerStarted","Data":"8b53228947c1ab7df276b647837b30ce72b300b72106f59cab51fab7e1228a98"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.263976 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:56 crc kubenswrapper[4824]: E1124 13:16:56.264350 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.764336056 +0000 UTC m=+158.403875366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.302385 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q"] Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.341318 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" event={"ID":"5efdc9d2-2541-4550-9f68-485fbc54473f","Type":"ContainerStarted","Data":"a50957309e1ca7e2586f58549f5e38d85cd6c444f303b293283f084ceed02b3b"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.341617 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn"] Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.356901 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" event={"ID":"090539a8-ec0a-4406-a781-2d45333291a1","Type":"ContainerStarted","Data":"49b1c637ec619033685e3c52d621b3c2b00130550764a0f50d10f3e00b009f0f"} Nov 24 13:16:56 crc kubenswrapper[4824]: E1124 13:16:56.370148 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.8701349 +0000 UTC m=+158.509674210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.369793 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.372125 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" event={"ID":"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6","Type":"ContainerStarted","Data":"1e6ee5d522a845a09ad7dbadbbf16fd9e2214f7ebff8b14b589d0ca56db4855b"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.388621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" event={"ID":"d63938dd-7154-4aac-9071-59a7803032c9","Type":"ContainerStarted","Data":"19304acf40b709af02e702b620c8336b960c002a069e818eabcc671e166468a6"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.409773 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" event={"ID":"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2","Type":"ContainerStarted","Data":"8e7cdd896d19e9d3374b2d843f20f5f7ae964ed50e309184c3536cf36e0169ee"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.432237 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" podStartSLOduration=130.432214912 podStartE2EDuration="2m10.432214912s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:56.430480708 +0000 UTC m=+158.070020018" watchObservedRunningTime="2025-11-24 13:16:56.432214912 +0000 UTC m=+158.071754222" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.447984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7lwbq" event={"ID":"3f3d1884-734b-4dd3-a795-356a8e561d7b","Type":"ContainerStarted","Data":"b8775a911803b858e7b82e6cd3bae4c4b246b0401aa9851a04fc05961d925184"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.454731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" event={"ID":"755c7bea-194d-4956-b24b-e2e65f031f38","Type":"ContainerStarted","Data":"4d1263e7aa9cc867f90dc0cca71e21a102dc6ee01b3b07a772fd029fccb5340d"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.464454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vmp98" event={"ID":"f457f2da-124f-43cd-ba58-1deddaff79d7","Type":"ContainerStarted","Data":"b824bf05a93f73ff16cac3141373b25231ec58f554f7fa759239ae8e6d56b9e8"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.471201 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:56 crc kubenswrapper[4824]: E1124 13:16:56.472263 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:56.972246772 +0000 UTC m=+158.611786082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.484709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" event={"ID":"72e37f91-f146-4005-a77f-dc7fba802ee1","Type":"ContainerStarted","Data":"f603f008e803125709214ddb63e1ac77adfe3efb400fa6225d32ee0e327126f0"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.504245 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" event={"ID":"7e7668d3-1c2b-4c90-9940-c1898a6c5069","Type":"ContainerStarted","Data":"f12ee348f2a5e1df12a4426209a9e052fc768c9b729ff7729ea52a81f7fbc50c"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.508287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn" event={"ID":"ecdd339c-197e-4525-8914-94be40499501","Type":"ContainerStarted","Data":"8d945b45a4e85345153995b1759d0c2a5569591175f221ab41b41a3c41623888"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.511957 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" event={"ID":"d41e0dcc-fb19-49fc-8aab-a1837193c058","Type":"ContainerStarted","Data":"078572340f12070746dca987cb02e7c4f6bc7178719db0dc2ec4c4513e7b6c1b"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.529119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" event={"ID":"a3bce575-3b95-4fb4-aec2-8a3461b330d5","Type":"ContainerStarted","Data":"9163bc0d3e6efaac7fda55984637fb0505ca267e5a21c8d203240083e92e99ae"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.533042 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" event={"ID":"2beaa442-848d-433a-841d-a461328b0376","Type":"ContainerStarted","Data":"697721b10203ed2166eea3fe24b0128a7307b85a551130b3277530249e1be722"} Nov 24 13:16:56 crc kubenswrapper[4824]: W1124 13:16:56.541188 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d01552e_05e5_4c7a_a32f_a1fcd31d748f.slice/crio-9e1aa47481f437c3d4de3925e4811494e68584f4d55e08658e8fa8a0b9e7eba4 WatchSource:0}: Error finding container 9e1aa47481f437c3d4de3925e4811494e68584f4d55e08658e8fa8a0b9e7eba4: Status 404 returned error can't find the container with id 9e1aa47481f437c3d4de3925e4811494e68584f4d55e08658e8fa8a0b9e7eba4 Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.541382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" event={"ID":"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0","Type":"ContainerStarted","Data":"92d359667845123bce77b0bd1e4d1d7d2dd0cc5c8f97ec47db5b0ea5945e2a4e"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.544338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" event={"ID":"48817d81-2ec0-4b16-b47c-b19e29b17b62","Type":"ContainerStarted","Data":"a8ecd0bf32e71bf66b5b7bdf919af99bb262969fd0f03e3726aad6f94dcf8e68"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.545015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" event={"ID":"cd954f01-b73c-4c86-bebf-59fa962e70fe","Type":"ContainerStarted","Data":"47461c3ff66ba38168d41aac8fda57b153619570ee78dbe77b86844cdcec33df"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.545840 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" event={"ID":"01ec77ac-5384-4457-8d8a-9d14135c2b82","Type":"ContainerStarted","Data":"a32f6638c06e20a97e68398a0b7cb3f470e852739cb40a0d105c0a88122c4288"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.550424 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-7lwbq" podStartSLOduration=131.550404625 podStartE2EDuration="2m11.550404625s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:56.484224381 +0000 UTC m=+158.123763691" watchObservedRunningTime="2025-11-24 13:16:56.550404625 +0000 UTC m=+158.189943935" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.551470 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxl9p" podStartSLOduration=131.551462662 podStartE2EDuration="2m11.551462662s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:56.549725408 +0000 UTC m=+158.189264718" watchObservedRunningTime="2025-11-24 13:16:56.551462662 +0000 UTC m=+158.191001972" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.564803 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" event={"ID":"3af944bb-d753-4fbe-a96c-20de1882117d","Type":"ContainerStarted","Data":"609e8aa49f794611252f60573b4ff26da49173e34d71410d936ff2d9512b1d79"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.575710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:56 crc kubenswrapper[4824]: E1124 13:16:56.577149 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.077136763 +0000 UTC m=+158.716676073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.614789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vwk9h" event={"ID":"ecea50ae-16db-46d7-a9cd-84a339185ae4","Type":"ContainerStarted","Data":"a10f4bef48b0c6a3048ce0b7ab6e91292a74d86e6f133537a1876f81696ce2e7"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.614843 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vwk9h" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.629945 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-vwk9h container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.629996 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vwk9h" podUID="ecea50ae-16db-46d7-a9cd-84a339185ae4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.633944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" event={"ID":"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054","Type":"ContainerStarted","Data":"1bb20ae0710dffbc2d364b4236ed5ec335f18d50a571b9602d1cf2205298a0f6"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.651663 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" podStartSLOduration=116.651647265 podStartE2EDuration="1m56.651647265s" podCreationTimestamp="2025-11-24 13:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:56.597107853 +0000 UTC m=+158.236647173" watchObservedRunningTime="2025-11-24 13:16:56.651647265 +0000 UTC m=+158.291186565" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.652160 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ll2hq" podStartSLOduration=131.652156768 podStartE2EDuration="2m11.652156768s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:56.651253686 +0000 UTC m=+158.290792996" watchObservedRunningTime="2025-11-24 13:16:56.652156768 +0000 UTC m=+158.291696078" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.683130 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:56 crc kubenswrapper[4824]: E1124 13:16:56.684233 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.184218189 +0000 UTC m=+158.823757489 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.699567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" event={"ID":"60b8f545-2446-4422-8838-ac6b54ea119f","Type":"ContainerStarted","Data":"afca24630d64297f92628fc0cfda6ee6a26240e952f0bf18ffb5c66e83a0759c"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.700423 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.701615 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.701656 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.708284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fnqzp" event={"ID":"002e2f07-c190-4749-aadf-3424ad02d9f0","Type":"ContainerStarted","Data":"70d4656e191ecaad76b7933b170797bc66da8e7895bf24e19c1451264f23bcd9"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.710292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" event={"ID":"7c508dba-0fa1-4f58-a939-061ff854c05c","Type":"ContainerStarted","Data":"fe4bba7371ac7d7226756fa7945e8630f2b9e5642697e23811a3a3d1d4c8ddc4"} Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.725279 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ss8j9" podStartSLOduration=131.725264655 podStartE2EDuration="2m11.725264655s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:56.682914617 +0000 UTC m=+158.322453927" watchObservedRunningTime="2025-11-24 13:16:56.725264655 +0000 UTC m=+158.364803965" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.726313 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vwk9h" podStartSLOduration=131.726306551 podStartE2EDuration="2m11.726306551s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:56.72503844 +0000 UTC m=+158.364577750" watchObservedRunningTime="2025-11-24 13:16:56.726306551 +0000 UTC m=+158.365845861" Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.784885 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:56 crc kubenswrapper[4824]: E1124 13:16:56.786683 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.28667157 +0000 UTC m=+158.926210880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:56 crc kubenswrapper[4824]: I1124 13:16:56.893879 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:56 crc kubenswrapper[4824]: E1124 13:16:56.895260 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.395240063 +0000 UTC m=+159.034779383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:56.996353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:56.996967 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.496955085 +0000 UTC m=+159.136494395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.097334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.097589 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.597560959 +0000 UTC m=+159.237100269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.097933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.098255 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.598245906 +0000 UTC m=+159.237785216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.200418 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.200676 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.700660206 +0000 UTC m=+159.340199516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.302192 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.302481 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.8024691 +0000 UTC m=+159.442008410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.380759 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.381143 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.403482 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.403894 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:57.903877564 +0000 UTC m=+159.543416874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.507318 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.507616 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.007601047 +0000 UTC m=+159.647140357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.608557 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.608941 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.108921769 +0000 UTC m=+159.748461089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.702227 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:16:57 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:16:57 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:16:57 crc kubenswrapper[4824]: healthz check failed Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.702272 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.712118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.712524 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.212509977 +0000 UTC m=+159.852049297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.719991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" event={"ID":"774e51b9-8e46-458f-82d2-e70027a8a0fc","Type":"ContainerStarted","Data":"3a3bcca2513da24c429c824268046f48b43cf9c89e7bf08ba763378b5419f14c"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.721990 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" event={"ID":"d41e0dcc-fb19-49fc-8aab-a1837193c058","Type":"ContainerStarted","Data":"9f824d869a0bb89d9f7516011d184f9e73e98c7582daf25b82e1ec9119e1088e"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.722690 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.723953 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ll2hq" event={"ID":"76c0d257-aca8-434e-97b2-067ddc1e7f66","Type":"ContainerStarted","Data":"0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.725165 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4fpg6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.44:8080/healthz\": dial tcp 10.217.0.44:8080: connect: connection refused" start-of-body= Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.725205 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" podUID="d41e0dcc-fb19-49fc-8aab-a1837193c058" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.44:8080/healthz\": dial tcp 10.217.0.44:8080: connect: connection refused" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.730501 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" event={"ID":"7c508dba-0fa1-4f58-a939-061ff854c05c","Type":"ContainerStarted","Data":"34f0d86af184535ddea3b972c349d4b45be210830e633bc6e910f1a7734e1680"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.734139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" event={"ID":"2cfbf6ab-805e-4776-8f3e-1823b675c1be","Type":"ContainerStarted","Data":"9e1e54bcf15ab8121d2b9372cd12b3e2e860ea5d38fc35fa9139623af99d7c67"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.745639 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" event={"ID":"090e8bea-a7c9-49cd-89dc-c2b0643e2bf0","Type":"ContainerStarted","Data":"882b7ef58e07e40b6a47d6eed22faf2ba9a791963511a6c6c89eb174f92b1019"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.747889 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7dczj" event={"ID":"9fb236a1-5efc-4aa7-9c19-692640304826","Type":"ContainerStarted","Data":"81ed2c49848dc1b939e506722ee12cb7d1924814ad29dee2ac2b298586b0f37d"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.752417 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" event={"ID":"91b65029-b851-4317-a1fc-26b30a5c8734","Type":"ContainerStarted","Data":"c9e9fb45eae69874c96a03cbc0745e19ddfe9e3fd2fe0c4a24683aeb8e2ac061"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.765336 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" event={"ID":"755c7bea-194d-4956-b24b-e2e65f031f38","Type":"ContainerStarted","Data":"be883bd2b5be4674cfb2eacd26417ef613f45ea0d60d2b31cef1256eaff19952"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.772298 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" event={"ID":"3d01552e-05e5-4c7a-a32f-a1fcd31d748f","Type":"ContainerStarted","Data":"9e1aa47481f437c3d4de3925e4811494e68584f4d55e08658e8fa8a0b9e7eba4"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.781311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" event={"ID":"adc0b256-e903-4e81-8d37-8f44d27234d1","Type":"ContainerStarted","Data":"dcfd587de76aecf267afc8bd4ad9c33fcae3f147327c6aab4d0f63dd82de5b3d"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.786620 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" event={"ID":"a3bce575-3b95-4fb4-aec2-8a3461b330d5","Type":"ContainerStarted","Data":"165c10b08c493e1ba20e44ec7bb00d46efb87a4b6100ca3c283404807b510895"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.787472 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.793268 4824 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-s96pq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.793453 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" podUID="a3bce575-3b95-4fb4-aec2-8a3461b330d5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.795019 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" event={"ID":"2beaa442-848d-433a-841d-a461328b0376","Type":"ContainerStarted","Data":"7b0f9a9418a7de9a058b534f4c86a03a4901fee55b51612f0f2510078874c67c"} Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.796664 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-vwk9h container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.796696 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vwk9h" podUID="ecea50ae-16db-46d7-a9cd-84a339185ae4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.813514 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.814608 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.314577118 +0000 UTC m=+159.954116428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.850659 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rk5nj" podStartSLOduration=132.850641849 podStartE2EDuration="2m12.850641849s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:57.849499041 +0000 UTC m=+159.489038351" watchObservedRunningTime="2025-11-24 13:16:57.850641849 +0000 UTC m=+159.490181159" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.880164 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-dw5hv" podStartSLOduration=132.880142657 podStartE2EDuration="2m12.880142657s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:57.877349867 +0000 UTC m=+159.516889197" watchObservedRunningTime="2025-11-24 13:16:57.880142657 +0000 UTC m=+159.519681967" Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.916513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:57 crc kubenswrapper[4824]: E1124 13:16:57.920421 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.420405243 +0000 UTC m=+160.059944553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:57 crc kubenswrapper[4824]: I1124 13:16:57.975025 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.018460 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.019159 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.51913963 +0000 UTC m=+160.158678940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.087729 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" podStartSLOduration=132.087712274 podStartE2EDuration="2m12.087712274s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:58.086505344 +0000 UTC m=+159.726044664" watchObservedRunningTime="2025-11-24 13:16:58.087712274 +0000 UTC m=+159.727251584" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.089678 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rdrk8" podStartSLOduration=132.089665493 podStartE2EDuration="2m12.089665493s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:57.971699915 +0000 UTC m=+159.611239225" watchObservedRunningTime="2025-11-24 13:16:58.089665493 +0000 UTC m=+159.729204803" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.120496 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.120855 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.620836292 +0000 UTC m=+160.260375602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.162246 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-j9k4z" podStartSLOduration=133.162226496 podStartE2EDuration="2m13.162226496s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:58.144272087 +0000 UTC m=+159.783811397" watchObservedRunningTime="2025-11-24 13:16:58.162226496 +0000 UTC m=+159.801765806" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.183361 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-7dczj" podStartSLOduration=8.183345464 podStartE2EDuration="8.183345464s" podCreationTimestamp="2025-11-24 13:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:58.18278115 +0000 UTC m=+159.822320460" watchObservedRunningTime="2025-11-24 13:16:58.183345464 +0000 UTC m=+159.822884774" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.227227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.227317 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.727303282 +0000 UTC m=+160.366842592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.227582 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.227878 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.727870797 +0000 UTC m=+160.367410107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.257001 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" podStartSLOduration=132.256982364 podStartE2EDuration="2m12.256982364s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:58.251365844 +0000 UTC m=+159.890905154" watchObservedRunningTime="2025-11-24 13:16:58.256982364 +0000 UTC m=+159.896521674" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.330127 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zp9xk" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.330490 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.330838 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.83082415 +0000 UTC m=+160.470363460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.432475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.432937 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:58.932926391 +0000 UTC m=+160.572465701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.534246 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.534652 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.034636823 +0000 UTC m=+160.674176133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.636848 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.637146 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.137134285 +0000 UTC m=+160.776673595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.701176 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:16:58 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:16:58 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:16:58 crc kubenswrapper[4824]: healthz check failed Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.701229 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.737704 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.737869 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.237845311 +0000 UTC m=+160.877384621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.737916 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.738274 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.238262942 +0000 UTC m=+160.877802252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.815470 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vmp98" event={"ID":"f457f2da-124f-43cd-ba58-1deddaff79d7","Type":"ContainerStarted","Data":"ad9160f190921fe14d6bdb5486a5aaf568d4c59afa7939aaa92018ef6420c537"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.832429 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" event={"ID":"72e37f91-f146-4005-a77f-dc7fba802ee1","Type":"ContainerStarted","Data":"4c88f85a9413fc57486016dbb50d35aff17ca7e5c9528bce01744e2edbc5af15"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.838745 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn" event={"ID":"ecdd339c-197e-4525-8914-94be40499501","Type":"ContainerStarted","Data":"324f2e18ccb8dc264882b18a9d2b1116df1baf67b7609e74b9f7102a12c7094a"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.839134 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.839278 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.339259256 +0000 UTC m=+160.978798566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.839475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.839885 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.339869571 +0000 UTC m=+160.979408881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.851839 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" event={"ID":"fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2","Type":"ContainerStarted","Data":"cbebcce8637f9b72f8037c50ded2a86e4c23268e0bc568ae5db228c8cc9ef265"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.852044 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.854760 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-sh749 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.854798 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" podUID="fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.861548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" event={"ID":"91b65029-b851-4317-a1fc-26b30a5c8734","Type":"ContainerStarted","Data":"0c6393e7e35b30c4409c5efbd78d0bb7a491e5ca071b573c5bb53c590d3cda17"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.864434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fnqzp" event={"ID":"002e2f07-c190-4749-aadf-3424ad02d9f0","Type":"ContainerStarted","Data":"d6dbca357f8f41572a781d0aacdb7c7036f012b23e9b5d022174c64eb6a8160d"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.872076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" event={"ID":"48817d81-2ec0-4b16-b47c-b19e29b17b62","Type":"ContainerStarted","Data":"42a105c3bad108a16d82a30d7329a7e4e44ced5228be4682148eab12cdbc5c24"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.875066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" event={"ID":"3af944bb-d753-4fbe-a96c-20de1882117d","Type":"ContainerStarted","Data":"ac5ae379bec652566504a3b20b3fc341b95e0b9468f4cd91ba82e0a8f18d9ab9"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.893246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" event={"ID":"ba6e7ecf-ad36-4dd9-bfce-e5c240c07054","Type":"ContainerStarted","Data":"65c3be15a6b4e6c9aaa3a9831815b6e8d87da1b5804b1854747ca790e20c414a"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.904946 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" event={"ID":"755c7bea-194d-4956-b24b-e2e65f031f38","Type":"ContainerStarted","Data":"bfed1b0a0fd9f4b4305f97bfd0e40a7fd4c798cd21a878b55ea95eacf02c86ab"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.921537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" event={"ID":"3d01552e-05e5-4c7a-a32f-a1fcd31d748f","Type":"ContainerStarted","Data":"a6fc28efa76ea487205c1765d5fdf02f7633d05ec3219758a0615d1e74b4df0a"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.932177 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" event={"ID":"7c508dba-0fa1-4f58-a939-061ff854c05c","Type":"ContainerStarted","Data":"d852246ca3a35c588da226328c2007ea088b1eb2bd29330506d5cbb526146b75"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.939282 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pbpx4" podStartSLOduration=133.939265855 podStartE2EDuration="2m13.939265855s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:58.882077036 +0000 UTC m=+160.521616346" watchObservedRunningTime="2025-11-24 13:16:58.939265855 +0000 UTC m=+160.578805165" Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.940044 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.940358 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.440345072 +0000 UTC m=+161.079884382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.940405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:58 crc kubenswrapper[4824]: E1124 13:16:58.940675 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.44066838 +0000 UTC m=+161.080207690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.957867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" event={"ID":"d63938dd-7154-4aac-9071-59a7803032c9","Type":"ContainerStarted","Data":"1537077d2a648ff9fcb7389ab5fbf5139a2ac74ff948eab9a96adfec813f2ae4"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.964247 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" event={"ID":"60b8f545-2446-4422-8838-ac6b54ea119f","Type":"ContainerStarted","Data":"e6bfce2dbdadcb2ff5cf6466a1cbec01b1365a7372c91bdb8ddd9f3dd480bfc8"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.994156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" event={"ID":"4cf5d5ca-477f-4750-8cd1-78a1db27f9d6","Type":"ContainerStarted","Data":"a8c4e685edd6a3527e913551fb41a7516906a5d6120827d276ce6d2edbf01d7e"} Nov 24 13:16:58 crc kubenswrapper[4824]: I1124 13:16:58.995248 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.002098 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zq9w9" podStartSLOduration=133.002078445 podStartE2EDuration="2m13.002078445s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:58.990705581 +0000 UTC m=+160.630244901" watchObservedRunningTime="2025-11-24 13:16:59.002078445 +0000 UTC m=+160.641617745" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.002188 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8h66q" podStartSLOduration=133.002184067 podStartE2EDuration="2m13.002184067s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:58.940206089 +0000 UTC m=+160.579745399" watchObservedRunningTime="2025-11-24 13:16:59.002184067 +0000 UTC m=+160.641723377" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.021026 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rl9nn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.021111 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" podUID="4cf5d5ca-477f-4750-8cd1-78a1db27f9d6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.040677 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4fpg6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.44:8080/healthz\": dial tcp 10.217.0.44:8080: connect: connection refused" start-of-body= Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.040717 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" podUID="d41e0dcc-fb19-49fc-8aab-a1837193c058" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.44:8080/healthz\": dial tcp 10.217.0.44:8080: connect: connection refused" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.042271 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.043371 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.543359556 +0000 UTC m=+161.182898866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.046194 4824 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-s96pq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.046237 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" podUID="a3bce575-3b95-4fb4-aec2-8a3461b330d5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.050487 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" event={"ID":"090539a8-ec0a-4406-a781-2d45333291a1","Type":"ContainerStarted","Data":"4161574676caacabd6bdbf337e7baa8b865e530f8245d4f4d66318ef000f513a"} Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.052883 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-857cb" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.114081 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ftftq" podStartSLOduration=133.114062543 podStartE2EDuration="2m13.114062543s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.073251984 +0000 UTC m=+160.712791294" watchObservedRunningTime="2025-11-24 13:16:59.114062543 +0000 UTC m=+160.753601853" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.114199 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-fnqzp" podStartSLOduration=9.114193097 podStartE2EDuration="9.114193097s" podCreationTimestamp="2025-11-24 13:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.106628518 +0000 UTC m=+160.746167838" watchObservedRunningTime="2025-11-24 13:16:59.114193097 +0000 UTC m=+160.753732407" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.145431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.168325 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.668305919 +0000 UTC m=+161.307845229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.169516 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" podStartSLOduration=133.169492239 podStartE2EDuration="2m13.169492239s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.166154345 +0000 UTC m=+160.805693645" watchObservedRunningTime="2025-11-24 13:16:59.169492239 +0000 UTC m=+160.809031549" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.230007 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nh8dl" podStartSLOduration=133.229991691 podStartE2EDuration="2m13.229991691s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.229092848 +0000 UTC m=+160.868632158" watchObservedRunningTime="2025-11-24 13:16:59.229991691 +0000 UTC m=+160.869531001" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.253343 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.253688 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.753669212 +0000 UTC m=+161.393208522 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.254083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.254474 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.754463542 +0000 UTC m=+161.394002852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.329415 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ps582" podStartSLOduration=133.329393625 podStartE2EDuration="2m13.329393625s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.283714093 +0000 UTC m=+160.923253403" watchObservedRunningTime="2025-11-24 13:16:59.329393625 +0000 UTC m=+160.968932935" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.355328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.355769 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.855751173 +0000 UTC m=+161.495290483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.437426 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" podStartSLOduration=133.437410204 podStartE2EDuration="2m13.437410204s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.394584084 +0000 UTC m=+161.034123394" watchObservedRunningTime="2025-11-24 13:16:59.437410204 +0000 UTC m=+161.076949514" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.437577 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kpqbs" podStartSLOduration=134.437573858 podStartE2EDuration="2m14.437573858s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.436764778 +0000 UTC m=+161.076304078" watchObservedRunningTime="2025-11-24 13:16:59.437573858 +0000 UTC m=+161.077113168" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.456877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.457323 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:16:59.95730231 +0000 UTC m=+161.596841650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.557716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.558136 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.05812049 +0000 UTC m=+161.697659800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.650929 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" podStartSLOduration=134.650909959 podStartE2EDuration="2m14.650909959s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.650366655 +0000 UTC m=+161.289905965" watchObservedRunningTime="2025-11-24 13:16:59.650909959 +0000 UTC m=+161.290449269" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.659674 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.660042 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.160025536 +0000 UTC m=+161.799564906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.675659 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xxpwx" podStartSLOduration=133.675643177 podStartE2EDuration="2m13.675643177s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:16:59.671431421 +0000 UTC m=+161.310970731" watchObservedRunningTime="2025-11-24 13:16:59.675643177 +0000 UTC m=+161.315182487" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.700310 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:16:59 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:16:59 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:16:59 crc kubenswrapper[4824]: healthz check failed Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.700543 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.761107 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.761496 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.261475922 +0000 UTC m=+161.901015232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.862937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.863327 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.363310057 +0000 UTC m=+162.002849367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.928568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.964212 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.964315 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.46429534 +0000 UTC m=+162.103834650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:16:59 crc kubenswrapper[4824]: I1124 13:16:59.964530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:16:59 crc kubenswrapper[4824]: E1124 13:16:59.964821 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.464795283 +0000 UTC m=+162.104334593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.035352 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vmp98" event={"ID":"f457f2da-124f-43cd-ba58-1deddaff79d7","Type":"ContainerStarted","Data":"2a8b104bda9bd79547816f3256f17c74edf9dde39b9385178fe0aabe96a80d66"} Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.036076 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vmp98" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.037081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn" event={"ID":"ecdd339c-197e-4525-8914-94be40499501","Type":"ContainerStarted","Data":"9aa3dcbe5dd60e153811b31f5e16b653d868631ae7dc30de7f60ac822e2c5be9"} Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.038820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" event={"ID":"3d01552e-05e5-4c7a-a32f-a1fcd31d748f","Type":"ContainerStarted","Data":"f9a95269df519cf33f93cbd63047480011a26ffaea6aadd45885e699523d36f6"} Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.039196 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.042400 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" event={"ID":"cd954f01-b73c-4c86-bebf-59fa962e70fe","Type":"ContainerStarted","Data":"ef3dc685be816aab896f086e79edecc4022eceafc7f0ba2f92897a9cffd9563b"} Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.042902 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4fpg6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.44:8080/healthz\": dial tcp 10.217.0.44:8080: connect: connection refused" start-of-body= Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.042935 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" podUID="d41e0dcc-fb19-49fc-8aab-a1837193c058" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.44:8080/healthz\": dial tcp 10.217.0.44:8080: connect: connection refused" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.043186 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rl9nn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.043204 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" podUID="4cf5d5ca-477f-4750-8cd1-78a1db27f9d6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.043581 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-sh749 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.043622 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" podUID="fb7ac5a4-915f-46e6-9ba0-8e6ce76423e2" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.043899 4824 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-s96pq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.043923 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" podUID="a3bce575-3b95-4fb4-aec2-8a3461b330d5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.065641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.066366 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.56633638 +0000 UTC m=+162.205875690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.097443 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vmp98" podStartSLOduration=10.097427587 podStartE2EDuration="10.097427587s" podCreationTimestamp="2025-11-24 13:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:17:00.070863864 +0000 UTC m=+161.710403174" watchObservedRunningTime="2025-11-24 13:17:00.097427587 +0000 UTC m=+161.736966897" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.123406 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" podStartSLOduration=134.123389826 podStartE2EDuration="2m14.123389826s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:17:00.116306809 +0000 UTC m=+161.755846109" watchObservedRunningTime="2025-11-24 13:17:00.123389826 +0000 UTC m=+161.762929136" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.159845 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rmdvn" podStartSLOduration=134.159825497 podStartE2EDuration="2m14.159825497s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:17:00.157753165 +0000 UTC m=+161.797292475" watchObservedRunningTime="2025-11-24 13:17:00.159825497 +0000 UTC m=+161.799364807" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.167144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.169343 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.669330934 +0000 UTC m=+162.308870244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.197077 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-mzl5h" podStartSLOduration=134.197059417 podStartE2EDuration="2m14.197059417s" podCreationTimestamp="2025-11-24 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:17:00.192428792 +0000 UTC m=+161.831968102" watchObservedRunningTime="2025-11-24 13:17:00.197059417 +0000 UTC m=+161.836598727" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.269026 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.269182 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.769155709 +0000 UTC m=+162.408695019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.269316 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.269637 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.769629331 +0000 UTC m=+162.409168631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.373141 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.373640 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.87362445 +0000 UTC m=+162.513163760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.475447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.475842 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:00.975826464 +0000 UTC m=+162.615365774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.576488 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.576913 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.07689433 +0000 UTC m=+162.716433640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.678596 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.679159 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.179145605 +0000 UTC m=+162.818684915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.701401 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:00 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:00 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:00 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.701454 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.780246 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.780568 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.280550089 +0000 UTC m=+162.920089399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.881844 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.882200 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.382185689 +0000 UTC m=+163.021724999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:00 crc kubenswrapper[4824]: I1124 13:17:00.983117 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:00 crc kubenswrapper[4824]: E1124 13:17:00.983499 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.483484071 +0000 UTC m=+163.123023381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.085007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.085416 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.585398798 +0000 UTC m=+163.224938108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.186221 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.186439 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.686407622 +0000 UTC m=+163.325946932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.186666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.187056 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.687043898 +0000 UTC m=+163.326583208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.288133 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.288548 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.788522964 +0000 UTC m=+163.428062274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.389487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.389825 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.889799345 +0000 UTC m=+163.529338655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.490174 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.490393 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.990364548 +0000 UTC m=+163.629903868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.490479 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.490748 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:01.990736757 +0000 UTC m=+163.630276067 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.592103 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.592653 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.092634944 +0000 UTC m=+163.732174244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.693678 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.694071 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.194056459 +0000 UTC m=+163.833595769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.701728 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:01 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:01 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:01 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.701792 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.794515 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.794845 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.294828657 +0000 UTC m=+163.934367957 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.824956 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.825006 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.896341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.897392 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.39737999 +0000 UTC m=+164.036919300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.961663 4824 patch_prober.go:28] interesting pod/apiserver-76f77b778f-t2fdq container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]log ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]etcd ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/generic-apiserver-start-informers ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/max-in-flight-filter ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 24 13:17:01 crc kubenswrapper[4824]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 24 13:17:01 crc kubenswrapper[4824]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/project.openshift.io-projectcache ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-startinformers ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 24 13:17:01 crc kubenswrapper[4824]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 24 13:17:01 crc kubenswrapper[4824]: livez check failed Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.961731 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" podUID="60b8f545-2446-4422-8838-ac6b54ea119f" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.997493 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.997664 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.497640125 +0000 UTC m=+164.137179435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:01 crc kubenswrapper[4824]: I1124 13:17:01.997791 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:01 crc kubenswrapper[4824]: E1124 13:17:01.998089 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.498074036 +0000 UTC m=+164.137613346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.047370 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rl9nn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.047432 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" podUID="4cf5d5ca-477f-4750-8cd1-78a1db27f9d6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.053038 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" event={"ID":"cd954f01-b73c-4c86-bebf-59fa962e70fe","Type":"ContainerStarted","Data":"a4c4d33314e0cedf442e23a78cadaa5d2d4612e79e76b521c196f413e2e88cf5"} Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.053095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" event={"ID":"cd954f01-b73c-4c86-bebf-59fa962e70fe","Type":"ContainerStarted","Data":"326e07afb79d2fc045081f857853773e56e5a54b0d7c5fab9efc0743215ca3f3"} Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.099591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.099735 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.599709346 +0000 UTC m=+164.239248656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.100276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.100751 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.600740582 +0000 UTC m=+164.240279892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.202497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.202824 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.702761631 +0000 UTC m=+164.342300941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.202984 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.203335 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.703316045 +0000 UTC m=+164.342855555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.304302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.304672 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.804638337 +0000 UTC m=+164.444177647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.305025 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.305434 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.805415317 +0000 UTC m=+164.444954627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.406394 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.406579 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.906553454 +0000 UTC m=+164.546092764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.406707 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.407097 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:02.907080388 +0000 UTC m=+164.546619698 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.508287 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.508480 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:03.008454811 +0000 UTC m=+164.647994121 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.508772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.509150 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:03.009129088 +0000 UTC m=+164.648668398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.609781 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.610013 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:03.109978128 +0000 UTC m=+164.749517438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.610839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.611315 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:03.111298631 +0000 UTC m=+164.750837941 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.626740 4824 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.702425 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:02 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:02 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:02 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.704076 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.712851 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.713206 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:03.213190668 +0000 UTC m=+164.852729978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.737323 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nv5fz"] Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.738283 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.742016 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.760562 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nv5fz"] Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.814625 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-catalog-content\") pod \"community-operators-nv5fz\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.814693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.814727 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k68n\" (UniqueName: \"kubernetes.io/projected/384f8f7c-bd08-4d8c-bf20-8680c8d49210-kube-api-access-4k68n\") pod \"community-operators-nv5fz\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.814759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-utilities\") pod \"community-operators-nv5fz\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.815082 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:03.315069034 +0000 UTC m=+164.954608334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.915380 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.915606 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-catalog-content\") pod \"community-operators-nv5fz\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.915666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k68n\" (UniqueName: \"kubernetes.io/projected/384f8f7c-bd08-4d8c-bf20-8680c8d49210-kube-api-access-4k68n\") pod \"community-operators-nv5fz\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.915695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-utilities\") pod \"community-operators-nv5fz\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.916209 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-utilities\") pod \"community-operators-nv5fz\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: E1124 13:17:02.916391 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:17:03.416368335 +0000 UTC m=+165.055907645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.916494 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-catalog-content\") pod \"community-operators-nv5fz\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.918964 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lfshk"] Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.920236 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.927022 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.942703 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k68n\" (UniqueName: \"kubernetes.io/projected/384f8f7c-bd08-4d8c-bf20-8680c8d49210-kube-api-access-4k68n\") pod \"community-operators-nv5fz\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:02 crc kubenswrapper[4824]: I1124 13:17:02.949253 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lfshk"] Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.017014 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.017079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lc8b\" (UniqueName: \"kubernetes.io/projected/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-kube-api-access-2lc8b\") pod \"certified-operators-lfshk\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.017142 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-utilities\") pod \"certified-operators-lfshk\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.017160 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-catalog-content\") pod \"certified-operators-lfshk\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: E1124 13:17:03.017436 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:17:03.517424751 +0000 UTC m=+165.156964061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zrmh8" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.045392 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-vwk9h container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.045443 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-vwk9h container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.045453 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vwk9h" podUID="ecea50ae-16db-46d7-a9cd-84a339185ae4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.045494 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vwk9h" podUID="ecea50ae-16db-46d7-a9cd-84a339185ae4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.049581 4824 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T13:17:02.626777288Z","Handler":null,"Name":""} Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.058452 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.061405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" event={"ID":"cd954f01-b73c-4c86-bebf-59fa962e70fe","Type":"ContainerStarted","Data":"1486b9a4dfb2ad214763abbe33e15b8c95d131faec3afd5e63ab5c79877705e2"} Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.090017 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-ksxcs" podStartSLOduration=13.090002273 podStartE2EDuration="13.090002273s" podCreationTimestamp="2025-11-24 13:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:17:03.089557392 +0000 UTC m=+164.729096722" watchObservedRunningTime="2025-11-24 13:17:03.090002273 +0000 UTC m=+164.729541583" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.113473 4824 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.114032 4824 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.118316 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.119011 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-catalog-content\") pod \"certified-operators-lfshk\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.119112 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-utilities\") pod \"certified-operators-lfshk\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.119221 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lc8b\" (UniqueName: \"kubernetes.io/projected/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-kube-api-access-2lc8b\") pod \"certified-operators-lfshk\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.119979 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-catalog-content\") pod \"certified-operators-lfshk\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.120176 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-utilities\") pod \"certified-operators-lfshk\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.124859 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.125021 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.125993 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cwsjh"] Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.126160 4824 patch_prober.go:28] interesting pod/console-f9d7485db-ll2hq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.126234 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ll2hq" podUID="76c0d257-aca8-434e-97b2-067ddc1e7f66" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.127336 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.138387 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.150693 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cwsjh"] Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.165764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lc8b\" (UniqueName: \"kubernetes.io/projected/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-kube-api-access-2lc8b\") pod \"certified-operators-lfshk\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.220118 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2nvn\" (UniqueName: \"kubernetes.io/projected/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-kube-api-access-c2nvn\") pod \"community-operators-cwsjh\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.220435 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-catalog-content\") pod \"community-operators-cwsjh\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.220618 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.220735 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-utilities\") pod \"community-operators-cwsjh\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.231772 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.232269 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.232203 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.290656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zrmh8\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.321722 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zpk2s"] Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.322832 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.326781 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2nvn\" (UniqueName: \"kubernetes.io/projected/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-kube-api-access-c2nvn\") pod \"community-operators-cwsjh\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.326862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-catalog-content\") pod \"community-operators-cwsjh\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.326924 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-utilities\") pod \"community-operators-cwsjh\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.327429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-utilities\") pod \"community-operators-cwsjh\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.327991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-catalog-content\") pod \"community-operators-cwsjh\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.362921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2nvn\" (UniqueName: \"kubernetes.io/projected/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-kube-api-access-c2nvn\") pod \"community-operators-cwsjh\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.374728 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpk2s"] Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.437248 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-catalog-content\") pod \"certified-operators-zpk2s\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.437340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-utilities\") pod \"certified-operators-zpk2s\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.437437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78x92\" (UniqueName: \"kubernetes.io/projected/51db319d-7582-4e5f-8b7e-b0f11a9e7757-kube-api-access-78x92\") pod \"certified-operators-zpk2s\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.446179 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.541471 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-utilities\") pod \"certified-operators-zpk2s\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.541905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78x92\" (UniqueName: \"kubernetes.io/projected/51db319d-7582-4e5f-8b7e-b0f11a9e7757-kube-api-access-78x92\") pod \"certified-operators-zpk2s\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.541949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-catalog-content\") pod \"certified-operators-zpk2s\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.542451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-catalog-content\") pod \"certified-operators-zpk2s\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.542628 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-utilities\") pod \"certified-operators-zpk2s\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.559144 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.569945 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78x92\" (UniqueName: \"kubernetes.io/projected/51db319d-7582-4e5f-8b7e-b0f11a9e7757-kube-api-access-78x92\") pod \"certified-operators-zpk2s\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.631378 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nv5fz"] Nov 24 13:17:03 crc kubenswrapper[4824]: W1124 13:17:03.637423 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod384f8f7c_bd08_4d8c_bf20_8680c8d49210.slice/crio-03a681a4ad208d06aa1aedee84264adc7576ac40d06e7db0164ef43e8137198c WatchSource:0}: Error finding container 03a681a4ad208d06aa1aedee84264adc7576ac40d06e7db0164ef43e8137198c: Status 404 returned error can't find the container with id 03a681a4ad208d06aa1aedee84264adc7576ac40d06e7db0164ef43e8137198c Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.689845 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.697160 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.704127 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:03 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:03 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:03 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.704189 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.744128 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s96pq" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.751706 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh749" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.805512 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.856225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl9nn" Nov 24 13:17:03 crc kubenswrapper[4824]: I1124 13:17:03.995184 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lfshk"] Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.050087 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpk2s"] Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.093119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpk2s" event={"ID":"51db319d-7582-4e5f-8b7e-b0f11a9e7757","Type":"ContainerStarted","Data":"14fc955284e01a931d2f1c1ce54336629ee633ad9f5987ccea9bf8ae594ee06d"} Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.108936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfshk" event={"ID":"a0b31fef-5fc1-4783-ba71-a02a9b01b21a","Type":"ContainerStarted","Data":"16920d1119b79b5b7c2b3b6086d0f37cfdc11eb109a9d326a42ca26f3cce16f1"} Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.113928 4824 generic.go:334] "Generic (PLEG): container finished" podID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerID="d75734caa2bb85f120c9fcffb68581dd189a00ba878cf1a93d9f9eb6dc0f45cf" exitCode=0 Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.115323 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv5fz" event={"ID":"384f8f7c-bd08-4d8c-bf20-8680c8d49210","Type":"ContainerDied","Data":"d75734caa2bb85f120c9fcffb68581dd189a00ba878cf1a93d9f9eb6dc0f45cf"} Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.115351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv5fz" event={"ID":"384f8f7c-bd08-4d8c-bf20-8680c8d49210","Type":"ContainerStarted","Data":"03a681a4ad208d06aa1aedee84264adc7576ac40d06e7db0164ef43e8137198c"} Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.121113 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.139258 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zrmh8"] Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.163125 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cwsjh"] Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.699929 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:04 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:04 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:04 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.699987 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.980296 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h5pzc"] Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.982241 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.982763 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.983188 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.996643 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.998127 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 13:17:04 crc kubenswrapper[4824]: I1124 13:17:04.998267 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.005351 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.019891 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.020731 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5pzc"] Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.033927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ac20781-a074-4984-8fd5-c756344632e2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7ac20781-a074-4984-8fd5-c756344632e2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.033984 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ac20781-a074-4984-8fd5-c756344632e2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7ac20781-a074-4984-8fd5-c756344632e2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.034010 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdnql\" (UniqueName: \"kubernetes.io/projected/163e9060-188a-41c6-8495-1e48b84ed3ba-kube-api-access-wdnql\") pod \"redhat-marketplace-h5pzc\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.034030 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-catalog-content\") pod \"redhat-marketplace-h5pzc\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.034092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-utilities\") pod \"redhat-marketplace-h5pzc\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.127056 4824 generic.go:334] "Generic (PLEG): container finished" podID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerID="e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248" exitCode=0 Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.127119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpk2s" event={"ID":"51db319d-7582-4e5f-8b7e-b0f11a9e7757","Type":"ContainerDied","Data":"e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248"} Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.129589 4824 generic.go:334] "Generic (PLEG): container finished" podID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerID="54a01f87877e9239f85b8e1029def5c38fb006c38828e3a2f55114e35ed5ab8a" exitCode=0 Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.129664 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfshk" event={"ID":"a0b31fef-5fc1-4783-ba71-a02a9b01b21a","Type":"ContainerDied","Data":"54a01f87877e9239f85b8e1029def5c38fb006c38828e3a2f55114e35ed5ab8a"} Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.135671 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-utilities\") pod \"redhat-marketplace-h5pzc\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.135743 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ac20781-a074-4984-8fd5-c756344632e2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7ac20781-a074-4984-8fd5-c756344632e2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.135771 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ac20781-a074-4984-8fd5-c756344632e2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7ac20781-a074-4984-8fd5-c756344632e2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.135800 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdnql\" (UniqueName: \"kubernetes.io/projected/163e9060-188a-41c6-8495-1e48b84ed3ba-kube-api-access-wdnql\") pod \"redhat-marketplace-h5pzc\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.135841 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-catalog-content\") pod \"redhat-marketplace-h5pzc\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.136286 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ac20781-a074-4984-8fd5-c756344632e2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7ac20781-a074-4984-8fd5-c756344632e2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.136485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-catalog-content\") pod \"redhat-marketplace-h5pzc\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.136594 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-utilities\") pod \"redhat-marketplace-h5pzc\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.144039 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" event={"ID":"b36ecad7-891d-4814-8a54-6a3c196c0525","Type":"ContainerStarted","Data":"4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf"} Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.144078 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" event={"ID":"b36ecad7-891d-4814-8a54-6a3c196c0525","Type":"ContainerStarted","Data":"26e510b7ce455c9aac0a85dff7baf65416a852bbf4b8662fb4dd849bd5bdccb6"} Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.144558 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.155432 4824 generic.go:334] "Generic (PLEG): container finished" podID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerID="9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e" exitCode=0 Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.155789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwsjh" event={"ID":"54a682b5-c0e0-4432-b69e-a98f0f94bd0c","Type":"ContainerDied","Data":"9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e"} Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.155838 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwsjh" event={"ID":"54a682b5-c0e0-4432-b69e-a98f0f94bd0c","Type":"ContainerStarted","Data":"4ee29cbe955a54ed7b00ee0adfc872e96cbb26fa76fdbccb0115929d9e046253"} Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.160632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdnql\" (UniqueName: \"kubernetes.io/projected/163e9060-188a-41c6-8495-1e48b84ed3ba-kube-api-access-wdnql\") pod \"redhat-marketplace-h5pzc\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.166765 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ac20781-a074-4984-8fd5-c756344632e2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7ac20781-a074-4984-8fd5-c756344632e2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.214514 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" podStartSLOduration=140.214493576 podStartE2EDuration="2m20.214493576s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:17:05.213141552 +0000 UTC m=+166.852680862" watchObservedRunningTime="2025-11-24 13:17:05.214493576 +0000 UTC m=+166.854032886" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.301369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.317679 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.338875 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6hvw9"] Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.340210 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.361789 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hvw9"] Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.449476 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-utilities\") pod \"redhat-marketplace-6hvw9\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.449529 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9cgc\" (UniqueName: \"kubernetes.io/projected/b8106bbc-8aca-4500-8ca9-38766f039302-kube-api-access-n9cgc\") pod \"redhat-marketplace-6hvw9\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.449649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-catalog-content\") pod \"redhat-marketplace-6hvw9\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.550759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9cgc\" (UniqueName: \"kubernetes.io/projected/b8106bbc-8aca-4500-8ca9-38766f039302-kube-api-access-n9cgc\") pod \"redhat-marketplace-6hvw9\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.550922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-catalog-content\") pod \"redhat-marketplace-6hvw9\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.550963 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-utilities\") pod \"redhat-marketplace-6hvw9\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.552541 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-utilities\") pod \"redhat-marketplace-6hvw9\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.552599 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-catalog-content\") pod \"redhat-marketplace-6hvw9\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.578197 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9cgc\" (UniqueName: \"kubernetes.io/projected/b8106bbc-8aca-4500-8ca9-38766f039302-kube-api-access-n9cgc\") pod \"redhat-marketplace-6hvw9\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.675254 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 13:17:05 crc kubenswrapper[4824]: W1124 13:17:05.685388 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7ac20781_a074_4984_8fd5_c756344632e2.slice/crio-7020aa53941b9c0975dfb9dcb03b5a58156fbe76ddc23cf3aaf4961252cac155 WatchSource:0}: Error finding container 7020aa53941b9c0975dfb9dcb03b5a58156fbe76ddc23cf3aaf4961252cac155: Status 404 returned error can't find the container with id 7020aa53941b9c0975dfb9dcb03b5a58156fbe76ddc23cf3aaf4961252cac155 Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.701573 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:05 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:05 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:05 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.701634 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.711175 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.916389 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p65h5"] Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.917629 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.919846 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.929453 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p65h5"] Nov 24 13:17:05 crc kubenswrapper[4824]: I1124 13:17:05.946321 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5pzc"] Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.032659 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hvw9"] Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.055927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-utilities\") pod \"redhat-operators-p65h5\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.055995 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-catalog-content\") pod \"redhat-operators-p65h5\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.056032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z78pw\" (UniqueName: \"kubernetes.io/projected/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-kube-api-access-z78pw\") pod \"redhat-operators-p65h5\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.157125 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-catalog-content\") pod \"redhat-operators-p65h5\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.157179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z78pw\" (UniqueName: \"kubernetes.io/projected/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-kube-api-access-z78pw\") pod \"redhat-operators-p65h5\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.157255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-utilities\") pod \"redhat-operators-p65h5\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.157698 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-utilities\") pod \"redhat-operators-p65h5\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.158008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-catalog-content\") pod \"redhat-operators-p65h5\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.192684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z78pw\" (UniqueName: \"kubernetes.io/projected/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-kube-api-access-z78pw\") pod \"redhat-operators-p65h5\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.198272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ac20781-a074-4984-8fd5-c756344632e2","Type":"ContainerStarted","Data":"7020aa53941b9c0975dfb9dcb03b5a58156fbe76ddc23cf3aaf4961252cac155"} Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.206253 4824 generic.go:334] "Generic (PLEG): container finished" podID="2beaa442-848d-433a-841d-a461328b0376" containerID="7b0f9a9418a7de9a058b534f4c86a03a4901fee55b51612f0f2510078874c67c" exitCode=0 Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.206318 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" event={"ID":"2beaa442-848d-433a-841d-a461328b0376","Type":"ContainerDied","Data":"7b0f9a9418a7de9a058b534f4c86a03a4901fee55b51612f0f2510078874c67c"} Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.215242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hvw9" event={"ID":"b8106bbc-8aca-4500-8ca9-38766f039302","Type":"ContainerStarted","Data":"49cb0547efd2902a4caa4e4c5d9165647c3752834e128b188b0e073fcc45f8bb"} Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.221022 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5pzc" event={"ID":"163e9060-188a-41c6-8495-1e48b84ed3ba","Type":"ContainerStarted","Data":"00b10ba4175bb86230657b3eefc81621cb568b0d8d20a4a5b06eb95cea8e6a1e"} Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.289037 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.314231 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vtmpf"] Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.315209 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.322329 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vtmpf"] Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.462083 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-utilities\") pod \"redhat-operators-vtmpf\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.462356 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx4v9\" (UniqueName: \"kubernetes.io/projected/3946423b-0ffd-4991-87df-908557e14ff7-kube-api-access-zx4v9\") pod \"redhat-operators-vtmpf\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.462432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-catalog-content\") pod \"redhat-operators-vtmpf\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.463651 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.464350 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.472332 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.473740 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.481694 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.564990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53333b0e-04cf-4626-83a6-d7d19e93685b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53333b0e-04cf-4626-83a6-d7d19e93685b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.565037 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53333b0e-04cf-4626-83a6-d7d19e93685b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53333b0e-04cf-4626-83a6-d7d19e93685b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.565109 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-utilities\") pod \"redhat-operators-vtmpf\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.565128 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx4v9\" (UniqueName: \"kubernetes.io/projected/3946423b-0ffd-4991-87df-908557e14ff7-kube-api-access-zx4v9\") pod \"redhat-operators-vtmpf\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.571260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-utilities\") pod \"redhat-operators-vtmpf\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.571374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-catalog-content\") pod \"redhat-operators-vtmpf\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.572122 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-catalog-content\") pod \"redhat-operators-vtmpf\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.596899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx4v9\" (UniqueName: \"kubernetes.io/projected/3946423b-0ffd-4991-87df-908557e14ff7-kube-api-access-zx4v9\") pod \"redhat-operators-vtmpf\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.649119 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.673360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53333b0e-04cf-4626-83a6-d7d19e93685b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53333b0e-04cf-4626-83a6-d7d19e93685b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.673405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53333b0e-04cf-4626-83a6-d7d19e93685b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53333b0e-04cf-4626-83a6-d7d19e93685b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.673513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53333b0e-04cf-4626-83a6-d7d19e93685b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53333b0e-04cf-4626-83a6-d7d19e93685b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.677657 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p65h5"] Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.695600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53333b0e-04cf-4626-83a6-d7d19e93685b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53333b0e-04cf-4626-83a6-d7d19e93685b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.700291 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:06 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:06 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:06 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.700375 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:06 crc kubenswrapper[4824]: W1124 13:17:06.741725 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fa2b5e3_3fc7_449f_b9b1_57ee8c575edf.slice/crio-681477b961afcffc08f3879139346f52204822dfa7b6391ddff8792e91188c46 WatchSource:0}: Error finding container 681477b961afcffc08f3879139346f52204822dfa7b6391ddff8792e91188c46: Status 404 returned error can't find the container with id 681477b961afcffc08f3879139346f52204822dfa7b6391ddff8792e91188c46 Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.803757 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.832076 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:17:06 crc kubenswrapper[4824]: I1124 13:17:06.839802 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-t2fdq" Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.130215 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vtmpf"] Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.282468 4824 generic.go:334] "Generic (PLEG): container finished" podID="7ac20781-a074-4984-8fd5-c756344632e2" containerID="cc83721cdb9eef2cde2f3317ee281af4fd6e78ab16c84d77660b87288438f956" exitCode=0 Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.282731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ac20781-a074-4984-8fd5-c756344632e2","Type":"ContainerDied","Data":"cc83721cdb9eef2cde2f3317ee281af4fd6e78ab16c84d77660b87288438f956"} Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.296748 4824 generic.go:334] "Generic (PLEG): container finished" podID="b8106bbc-8aca-4500-8ca9-38766f039302" containerID="30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36" exitCode=0 Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.296843 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hvw9" event={"ID":"b8106bbc-8aca-4500-8ca9-38766f039302","Type":"ContainerDied","Data":"30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36"} Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.331988 4824 generic.go:334] "Generic (PLEG): container finished" podID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerID="51cc9bb4c19d428c34ea8d09e9d18b5ec806fde1276e6b21fba036127795aebb" exitCode=0 Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.332074 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5pzc" event={"ID":"163e9060-188a-41c6-8495-1e48b84ed3ba","Type":"ContainerDied","Data":"51cc9bb4c19d428c34ea8d09e9d18b5ec806fde1276e6b21fba036127795aebb"} Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.388427 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p65h5" event={"ID":"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf","Type":"ContainerStarted","Data":"d9de437acb26e2a4a56e4766b8877a424a529e5300495dae256bcaa63c2c7b25"} Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.388467 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p65h5" event={"ID":"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf","Type":"ContainerStarted","Data":"681477b961afcffc08f3879139346f52204822dfa7b6391ddff8792e91188c46"} Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.400202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtmpf" event={"ID":"3946423b-0ffd-4991-87df-908557e14ff7","Type":"ContainerStarted","Data":"6f6922125f7f826dc5dbc1f05d43589551832c53e3ebefa13f4674d8092f6d78"} Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.471452 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.708136 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:07 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:07 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:07 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.708194 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:07 crc kubenswrapper[4824]: I1124 13:17:07.862839 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.009191 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzvxc\" (UniqueName: \"kubernetes.io/projected/2beaa442-848d-433a-841d-a461328b0376-kube-api-access-jzvxc\") pod \"2beaa442-848d-433a-841d-a461328b0376\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.009275 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beaa442-848d-433a-841d-a461328b0376-secret-volume\") pod \"2beaa442-848d-433a-841d-a461328b0376\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.009390 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beaa442-848d-433a-841d-a461328b0376-config-volume\") pod \"2beaa442-848d-433a-841d-a461328b0376\" (UID: \"2beaa442-848d-433a-841d-a461328b0376\") " Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.010465 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2beaa442-848d-433a-841d-a461328b0376-config-volume" (OuterVolumeSpecName: "config-volume") pod "2beaa442-848d-433a-841d-a461328b0376" (UID: "2beaa442-848d-433a-841d-a461328b0376"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.029088 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2beaa442-848d-433a-841d-a461328b0376-kube-api-access-jzvxc" (OuterVolumeSpecName: "kube-api-access-jzvxc") pod "2beaa442-848d-433a-841d-a461328b0376" (UID: "2beaa442-848d-433a-841d-a461328b0376"). InnerVolumeSpecName "kube-api-access-jzvxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.030271 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2beaa442-848d-433a-841d-a461328b0376-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2beaa442-848d-433a-841d-a461328b0376" (UID: "2beaa442-848d-433a-841d-a461328b0376"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.111459 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beaa442-848d-433a-841d-a461328b0376-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.111489 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzvxc\" (UniqueName: \"kubernetes.io/projected/2beaa442-848d-433a-841d-a461328b0376-kube-api-access-jzvxc\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.111501 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beaa442-848d-433a-841d-a461328b0376-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.417870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" event={"ID":"2beaa442-848d-433a-841d-a461328b0376","Type":"ContainerDied","Data":"697721b10203ed2166eea3fe24b0128a7307b85a551130b3277530249e1be722"} Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.417907 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="697721b10203ed2166eea3fe24b0128a7307b85a551130b3277530249e1be722" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.417909 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.425009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53333b0e-04cf-4626-83a6-d7d19e93685b","Type":"ContainerStarted","Data":"4b99ab122188a5c745dcda95b09702a2f6468f7bdeb93c586b5494d2d7bf3768"} Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.425086 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53333b0e-04cf-4626-83a6-d7d19e93685b","Type":"ContainerStarted","Data":"48128f445517f63762476791b506222f403124275b85905d47cfacd50eda4552"} Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.429669 4824 generic.go:334] "Generic (PLEG): container finished" podID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerID="d9de437acb26e2a4a56e4766b8877a424a529e5300495dae256bcaa63c2c7b25" exitCode=0 Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.429741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p65h5" event={"ID":"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf","Type":"ContainerDied","Data":"d9de437acb26e2a4a56e4766b8877a424a529e5300495dae256bcaa63c2c7b25"} Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.445176 4824 generic.go:334] "Generic (PLEG): container finished" podID="3946423b-0ffd-4991-87df-908557e14ff7" containerID="7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246" exitCode=0 Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.445384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtmpf" event={"ID":"3946423b-0ffd-4991-87df-908557e14ff7","Type":"ContainerDied","Data":"7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246"} Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.445816 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.445794358 podStartE2EDuration="2.445794358s" podCreationTimestamp="2025-11-24 13:17:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:17:08.444464175 +0000 UTC m=+170.084003485" watchObservedRunningTime="2025-11-24 13:17:08.445794358 +0000 UTC m=+170.085333668" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.618531 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.650212 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7e81fafe-6567-47f5-bb87-78fffb25fa9b-metrics-certs\") pod \"network-metrics-daemon-sc887\" (UID: \"7e81fafe-6567-47f5-bb87-78fffb25fa9b\") " pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.701284 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:08 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:08 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:08 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.701354 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.730903 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sc887" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.824542 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.864353 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vmp98" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.925494 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ac20781-a074-4984-8fd5-c756344632e2-kube-api-access\") pod \"7ac20781-a074-4984-8fd5-c756344632e2\" (UID: \"7ac20781-a074-4984-8fd5-c756344632e2\") " Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.925764 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ac20781-a074-4984-8fd5-c756344632e2-kubelet-dir\") pod \"7ac20781-a074-4984-8fd5-c756344632e2\" (UID: \"7ac20781-a074-4984-8fd5-c756344632e2\") " Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.927945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ac20781-a074-4984-8fd5-c756344632e2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7ac20781-a074-4984-8fd5-c756344632e2" (UID: "7ac20781-a074-4984-8fd5-c756344632e2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.930507 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ac20781-a074-4984-8fd5-c756344632e2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:08 crc kubenswrapper[4824]: I1124 13:17:08.933005 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac20781-a074-4984-8fd5-c756344632e2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7ac20781-a074-4984-8fd5-c756344632e2" (UID: "7ac20781-a074-4984-8fd5-c756344632e2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.031214 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ac20781-a074-4984-8fd5-c756344632e2-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.250951 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-sc887"] Nov 24 13:17:09 crc kubenswrapper[4824]: W1124 13:17:09.290658 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e81fafe_6567_47f5_bb87_78fffb25fa9b.slice/crio-8cc45ca915935d4148cf2cb6d2cd0ef57abfcb54fbafefb58440a36e04135ff4 WatchSource:0}: Error finding container 8cc45ca915935d4148cf2cb6d2cd0ef57abfcb54fbafefb58440a36e04135ff4: Status 404 returned error can't find the container with id 8cc45ca915935d4148cf2cb6d2cd0ef57abfcb54fbafefb58440a36e04135ff4 Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.507718 4824 generic.go:334] "Generic (PLEG): container finished" podID="53333b0e-04cf-4626-83a6-d7d19e93685b" containerID="4b99ab122188a5c745dcda95b09702a2f6468f7bdeb93c586b5494d2d7bf3768" exitCode=0 Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.507819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53333b0e-04cf-4626-83a6-d7d19e93685b","Type":"ContainerDied","Data":"4b99ab122188a5c745dcda95b09702a2f6468f7bdeb93c586b5494d2d7bf3768"} Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.533983 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.534326 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ac20781-a074-4984-8fd5-c756344632e2","Type":"ContainerDied","Data":"7020aa53941b9c0975dfb9dcb03b5a58156fbe76ddc23cf3aaf4961252cac155"} Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.534397 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7020aa53941b9c0975dfb9dcb03b5a58156fbe76ddc23cf3aaf4961252cac155" Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.546127 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-sc887" event={"ID":"7e81fafe-6567-47f5-bb87-78fffb25fa9b","Type":"ContainerStarted","Data":"8cc45ca915935d4148cf2cb6d2cd0ef57abfcb54fbafefb58440a36e04135ff4"} Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.705039 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:09 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:09 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:09 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:09 crc kubenswrapper[4824]: I1124 13:17:09.705435 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:10 crc kubenswrapper[4824]: I1124 13:17:10.590090 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-sc887" event={"ID":"7e81fafe-6567-47f5-bb87-78fffb25fa9b","Type":"ContainerStarted","Data":"f334e5f2e0a3a9ac1df144a5376d19db7a99ac5842fa2a36077349013245a04a"} Nov 24 13:17:10 crc kubenswrapper[4824]: I1124 13:17:10.701477 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:10 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:10 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:10 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:10 crc kubenswrapper[4824]: I1124 13:17:10.701533 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:10 crc kubenswrapper[4824]: I1124 13:17:10.788296 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:17:10 crc kubenswrapper[4824]: I1124 13:17:10.788697 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.001214 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.055038 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53333b0e-04cf-4626-83a6-d7d19e93685b-kube-api-access\") pod \"53333b0e-04cf-4626-83a6-d7d19e93685b\" (UID: \"53333b0e-04cf-4626-83a6-d7d19e93685b\") " Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.055704 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53333b0e-04cf-4626-83a6-d7d19e93685b-kubelet-dir\") pod \"53333b0e-04cf-4626-83a6-d7d19e93685b\" (UID: \"53333b0e-04cf-4626-83a6-d7d19e93685b\") " Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.055901 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53333b0e-04cf-4626-83a6-d7d19e93685b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "53333b0e-04cf-4626-83a6-d7d19e93685b" (UID: "53333b0e-04cf-4626-83a6-d7d19e93685b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.056598 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53333b0e-04cf-4626-83a6-d7d19e93685b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.065072 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53333b0e-04cf-4626-83a6-d7d19e93685b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "53333b0e-04cf-4626-83a6-d7d19e93685b" (UID: "53333b0e-04cf-4626-83a6-d7d19e93685b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.157789 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53333b0e-04cf-4626-83a6-d7d19e93685b-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.657262 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-sc887" event={"ID":"7e81fafe-6567-47f5-bb87-78fffb25fa9b","Type":"ContainerStarted","Data":"861e11c7ef88db0ba810e56c8f3818371ff41af65f52a8c063f16aee1f4a38ff"} Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.702287 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:11 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:11 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:11 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.702400 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.713002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53333b0e-04cf-4626-83a6-d7d19e93685b","Type":"ContainerDied","Data":"48128f445517f63762476791b506222f403124275b85905d47cfacd50eda4552"} Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.713052 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48128f445517f63762476791b506222f403124275b85905d47cfacd50eda4552" Nov 24 13:17:11 crc kubenswrapper[4824]: I1124 13:17:11.713232 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:17:12 crc kubenswrapper[4824]: I1124 13:17:12.699754 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:12 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:12 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:12 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:12 crc kubenswrapper[4824]: I1124 13:17:12.699863 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:13 crc kubenswrapper[4824]: I1124 13:17:13.045381 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-vwk9h container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 24 13:17:13 crc kubenswrapper[4824]: I1124 13:17:13.045454 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vwk9h" podUID="ecea50ae-16db-46d7-a9cd-84a339185ae4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 24 13:17:13 crc kubenswrapper[4824]: I1124 13:17:13.045463 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-vwk9h container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Nov 24 13:17:13 crc kubenswrapper[4824]: I1124 13:17:13.045539 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vwk9h" podUID="ecea50ae-16db-46d7-a9cd-84a339185ae4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Nov 24 13:17:13 crc kubenswrapper[4824]: I1124 13:17:13.125050 4824 patch_prober.go:28] interesting pod/console-f9d7485db-ll2hq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 24 13:17:13 crc kubenswrapper[4824]: I1124 13:17:13.125105 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ll2hq" podUID="76c0d257-aca8-434e-97b2-067ddc1e7f66" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 24 13:17:13 crc kubenswrapper[4824]: I1124 13:17:13.700409 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7lwbq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:17:13 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Nov 24 13:17:13 crc kubenswrapper[4824]: [+]process-running ok Nov 24 13:17:13 crc kubenswrapper[4824]: healthz check failed Nov 24 13:17:13 crc kubenswrapper[4824]: I1124 13:17:13.700558 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7lwbq" podUID="3f3d1884-734b-4dd3-a795-356a8e561d7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:17:14 crc kubenswrapper[4824]: I1124 13:17:14.700647 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:17:14 crc kubenswrapper[4824]: I1124 13:17:14.704549 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-7lwbq" Nov 24 13:17:14 crc kubenswrapper[4824]: I1124 13:17:14.723084 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-sc887" podStartSLOduration=149.72306085 podStartE2EDuration="2m29.72306085s" podCreationTimestamp="2025-11-24 13:14:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:17:11.678958946 +0000 UTC m=+173.318498286" watchObservedRunningTime="2025-11-24 13:17:14.72306085 +0000 UTC m=+176.362600170" Nov 24 13:17:18 crc kubenswrapper[4824]: I1124 13:17:18.150160 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:17:23 crc kubenswrapper[4824]: I1124 13:17:23.050230 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vwk9h" Nov 24 13:17:23 crc kubenswrapper[4824]: I1124 13:17:23.129326 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:17:23 crc kubenswrapper[4824]: I1124 13:17:23.133649 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:17:23 crc kubenswrapper[4824]: I1124 13:17:23.568257 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:17:34 crc kubenswrapper[4824]: I1124 13:17:34.082365 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ml8pn" Nov 24 13:17:40 crc kubenswrapper[4824]: I1124 13:17:40.788330 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:17:40 crc kubenswrapper[4824]: I1124 13:17:40.788898 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:17:41 crc kubenswrapper[4824]: E1124 13:17:41.953259 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 13:17:41 crc kubenswrapper[4824]: E1124 13:17:41.953445 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-78x92,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zpk2s_openshift-marketplace(51db319d-7582-4e5f-8b7e-b0f11a9e7757): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:17:41 crc kubenswrapper[4824]: E1124 13:17:41.954898 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zpk2s" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" Nov 24 13:17:42 crc kubenswrapper[4824]: E1124 13:17:42.028657 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 13:17:42 crc kubenswrapper[4824]: E1124 13:17:42.028830 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2lc8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-lfshk_openshift-marketplace(a0b31fef-5fc1-4783-ba71-a02a9b01b21a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:17:42 crc kubenswrapper[4824]: E1124 13:17:42.029917 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-lfshk" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" Nov 24 13:17:42 crc kubenswrapper[4824]: E1124 13:17:42.037752 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 13:17:42 crc kubenswrapper[4824]: E1124 13:17:42.037921 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c2nvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cwsjh_openshift-marketplace(54a682b5-c0e0-4432-b69e-a98f0f94bd0c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:17:42 crc kubenswrapper[4824]: E1124 13:17:42.039503 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cwsjh" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.945985 4824 generic.go:334] "Generic (PLEG): container finished" podID="b8106bbc-8aca-4500-8ca9-38766f039302" containerID="6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72" exitCode=0 Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.946021 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hvw9" event={"ID":"b8106bbc-8aca-4500-8ca9-38766f039302","Type":"ContainerDied","Data":"6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72"} Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.950422 4824 generic.go:334] "Generic (PLEG): container finished" podID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerID="d5b61120056d86f2b996f2b97208d28f17c0d2a4bcaa1bbcdebd48342e7fa1b0" exitCode=0 Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.950502 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5pzc" event={"ID":"163e9060-188a-41c6-8495-1e48b84ed3ba","Type":"ContainerDied","Data":"d5b61120056d86f2b996f2b97208d28f17c0d2a4bcaa1bbcdebd48342e7fa1b0"} Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.954989 4824 generic.go:334] "Generic (PLEG): container finished" podID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerID="a1aa18757dd55cef332f0d08956c44e369f85bf46f528ae1df35e98903efc852" exitCode=0 Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.955087 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p65h5" event={"ID":"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf","Type":"ContainerDied","Data":"a1aa18757dd55cef332f0d08956c44e369f85bf46f528ae1df35e98903efc852"} Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.959193 4824 generic.go:334] "Generic (PLEG): container finished" podID="3946423b-0ffd-4991-87df-908557e14ff7" containerID="743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368" exitCode=0 Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.959445 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtmpf" event={"ID":"3946423b-0ffd-4991-87df-908557e14ff7","Type":"ContainerDied","Data":"743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368"} Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.972707 4824 generic.go:334] "Generic (PLEG): container finished" podID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerID="9269dc32b8598c028d9631dfc6e29e4d9b10902cde840627fe8312b17e68dc6a" exitCode=0 Nov 24 13:17:42 crc kubenswrapper[4824]: I1124 13:17:42.972822 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv5fz" event={"ID":"384f8f7c-bd08-4d8c-bf20-8680c8d49210","Type":"ContainerDied","Data":"9269dc32b8598c028d9631dfc6e29e4d9b10902cde840627fe8312b17e68dc6a"} Nov 24 13:17:42 crc kubenswrapper[4824]: E1124 13:17:42.976194 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-lfshk" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" Nov 24 13:17:42 crc kubenswrapper[4824]: E1124 13:17:42.979039 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cwsjh" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" Nov 24 13:17:42 crc kubenswrapper[4824]: E1124 13:17:42.979901 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zpk2s" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" Nov 24 13:17:45 crc kubenswrapper[4824]: I1124 13:17:45.986394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv5fz" event={"ID":"384f8f7c-bd08-4d8c-bf20-8680c8d49210","Type":"ContainerStarted","Data":"72712b195f31bb64170f683d28255ef1b3a96edfbda5ea3d681775dd9d3a87d1"} Nov 24 13:17:45 crc kubenswrapper[4824]: I1124 13:17:45.989548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hvw9" event={"ID":"b8106bbc-8aca-4500-8ca9-38766f039302","Type":"ContainerStarted","Data":"ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66"} Nov 24 13:17:45 crc kubenswrapper[4824]: I1124 13:17:45.991272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5pzc" event={"ID":"163e9060-188a-41c6-8495-1e48b84ed3ba","Type":"ContainerStarted","Data":"6c263c5d9cd3be467e651adba3bca87f57f2b2c3e3142647ef651bacdad15a4d"} Nov 24 13:17:45 crc kubenswrapper[4824]: I1124 13:17:45.993136 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p65h5" event={"ID":"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf","Type":"ContainerStarted","Data":"67364a06df60373f7864dc0404c368c30275c6820b6630850035292c1d2817a7"} Nov 24 13:17:45 crc kubenswrapper[4824]: I1124 13:17:45.994964 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtmpf" event={"ID":"3946423b-0ffd-4991-87df-908557e14ff7","Type":"ContainerStarted","Data":"6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d"} Nov 24 13:17:46 crc kubenswrapper[4824]: I1124 13:17:46.007008 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nv5fz" podStartSLOduration=3.096051148 podStartE2EDuration="44.006987099s" podCreationTimestamp="2025-11-24 13:17:02 +0000 UTC" firstStartedPulling="2025-11-24 13:17:04.120836505 +0000 UTC m=+165.760375815" lastFinishedPulling="2025-11-24 13:17:45.031772456 +0000 UTC m=+206.671311766" observedRunningTime="2025-11-24 13:17:46.004373272 +0000 UTC m=+207.643912582" watchObservedRunningTime="2025-11-24 13:17:46.006987099 +0000 UTC m=+207.646526419" Nov 24 13:17:46 crc kubenswrapper[4824]: I1124 13:17:46.028640 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h5pzc" podStartSLOduration=4.362553488 podStartE2EDuration="42.028619093s" podCreationTimestamp="2025-11-24 13:17:04 +0000 UTC" firstStartedPulling="2025-11-24 13:17:07.343260295 +0000 UTC m=+168.982799605" lastFinishedPulling="2025-11-24 13:17:45.0093259 +0000 UTC m=+206.648865210" observedRunningTime="2025-11-24 13:17:46.025443002 +0000 UTC m=+207.664982312" watchObservedRunningTime="2025-11-24 13:17:46.028619093 +0000 UTC m=+207.668158403" Nov 24 13:17:46 crc kubenswrapper[4824]: I1124 13:17:46.069058 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vtmpf" podStartSLOduration=3.388887327 podStartE2EDuration="40.06903665s" podCreationTimestamp="2025-11-24 13:17:06 +0000 UTC" firstStartedPulling="2025-11-24 13:17:08.447777358 +0000 UTC m=+170.087316668" lastFinishedPulling="2025-11-24 13:17:45.127926681 +0000 UTC m=+206.767465991" observedRunningTime="2025-11-24 13:17:46.048262127 +0000 UTC m=+207.687801447" watchObservedRunningTime="2025-11-24 13:17:46.06903665 +0000 UTC m=+207.708575960" Nov 24 13:17:46 crc kubenswrapper[4824]: I1124 13:17:46.070002 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p65h5" podStartSLOduration=3.229732188 podStartE2EDuration="41.069993824s" podCreationTimestamp="2025-11-24 13:17:05 +0000 UTC" firstStartedPulling="2025-11-24 13:17:07.395734756 +0000 UTC m=+169.035274056" lastFinishedPulling="2025-11-24 13:17:45.235996382 +0000 UTC m=+206.875535692" observedRunningTime="2025-11-24 13:17:46.066365321 +0000 UTC m=+207.705904631" watchObservedRunningTime="2025-11-24 13:17:46.069993824 +0000 UTC m=+207.709533134" Nov 24 13:17:46 crc kubenswrapper[4824]: I1124 13:17:46.096113 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6hvw9" podStartSLOduration=3.054459969 podStartE2EDuration="41.096091313s" podCreationTimestamp="2025-11-24 13:17:05 +0000 UTC" firstStartedPulling="2025-11-24 13:17:07.327583743 +0000 UTC m=+168.967123053" lastFinishedPulling="2025-11-24 13:17:45.369215087 +0000 UTC m=+207.008754397" observedRunningTime="2025-11-24 13:17:46.094069611 +0000 UTC m=+207.733608941" watchObservedRunningTime="2025-11-24 13:17:46.096091313 +0000 UTC m=+207.735630633" Nov 24 13:17:46 crc kubenswrapper[4824]: I1124 13:17:46.292662 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:46 crc kubenswrapper[4824]: I1124 13:17:46.292713 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:46 crc kubenswrapper[4824]: I1124 13:17:46.650133 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:46 crc kubenswrapper[4824]: I1124 13:17:46.650491 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:47 crc kubenswrapper[4824]: I1124 13:17:47.564663 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p65h5" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerName="registry-server" probeResult="failure" output=< Nov 24 13:17:47 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 13:17:47 crc kubenswrapper[4824]: > Nov 24 13:17:47 crc kubenswrapper[4824]: I1124 13:17:47.692933 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vtmpf" podUID="3946423b-0ffd-4991-87df-908557e14ff7" containerName="registry-server" probeResult="failure" output=< Nov 24 13:17:47 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 13:17:47 crc kubenswrapper[4824]: > Nov 24 13:17:53 crc kubenswrapper[4824]: I1124 13:17:53.059403 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:53 crc kubenswrapper[4824]: I1124 13:17:53.059864 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:53 crc kubenswrapper[4824]: I1124 13:17:53.109414 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:54 crc kubenswrapper[4824]: I1124 13:17:54.085030 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:17:55 crc kubenswrapper[4824]: I1124 13:17:55.318596 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:55 crc kubenswrapper[4824]: I1124 13:17:55.318637 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:55 crc kubenswrapper[4824]: I1124 13:17:55.354842 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:55 crc kubenswrapper[4824]: I1124 13:17:55.712595 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:55 crc kubenswrapper[4824]: I1124 13:17:55.712647 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:55 crc kubenswrapper[4824]: I1124 13:17:55.750297 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:56 crc kubenswrapper[4824]: I1124 13:17:56.084987 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:56 crc kubenswrapper[4824]: I1124 13:17:56.089792 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:17:56 crc kubenswrapper[4824]: I1124 13:17:56.339838 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:56 crc kubenswrapper[4824]: I1124 13:17:56.385956 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:17:56 crc kubenswrapper[4824]: I1124 13:17:56.687820 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:56 crc kubenswrapper[4824]: I1124 13:17:56.729172 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:56 crc kubenswrapper[4824]: I1124 13:17:56.782220 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hvw9"] Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.060538 4824 generic.go:334] "Generic (PLEG): container finished" podID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerID="4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4" exitCode=0 Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.061103 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6hvw9" podUID="b8106bbc-8aca-4500-8ca9-38766f039302" containerName="registry-server" containerID="cri-o://ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66" gracePeriod=2 Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.060732 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpk2s" event={"ID":"51db319d-7582-4e5f-8b7e-b0f11a9e7757","Type":"ContainerDied","Data":"4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4"} Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.422462 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.460611 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-catalog-content\") pod \"b8106bbc-8aca-4500-8ca9-38766f039302\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.460678 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-utilities\") pod \"b8106bbc-8aca-4500-8ca9-38766f039302\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.460793 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9cgc\" (UniqueName: \"kubernetes.io/projected/b8106bbc-8aca-4500-8ca9-38766f039302-kube-api-access-n9cgc\") pod \"b8106bbc-8aca-4500-8ca9-38766f039302\" (UID: \"b8106bbc-8aca-4500-8ca9-38766f039302\") " Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.461474 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-utilities" (OuterVolumeSpecName: "utilities") pod "b8106bbc-8aca-4500-8ca9-38766f039302" (UID: "b8106bbc-8aca-4500-8ca9-38766f039302"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.467122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8106bbc-8aca-4500-8ca9-38766f039302-kube-api-access-n9cgc" (OuterVolumeSpecName: "kube-api-access-n9cgc") pod "b8106bbc-8aca-4500-8ca9-38766f039302" (UID: "b8106bbc-8aca-4500-8ca9-38766f039302"). InnerVolumeSpecName "kube-api-access-n9cgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.479510 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8106bbc-8aca-4500-8ca9-38766f039302" (UID: "b8106bbc-8aca-4500-8ca9-38766f039302"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.562683 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9cgc\" (UniqueName: \"kubernetes.io/projected/b8106bbc-8aca-4500-8ca9-38766f039302-kube-api-access-n9cgc\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.562724 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.562738 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8106bbc-8aca-4500-8ca9-38766f039302-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.583341 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vtmpf"] Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.583922 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vtmpf" podUID="3946423b-0ffd-4991-87df-908557e14ff7" containerName="registry-server" containerID="cri-o://6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d" gracePeriod=2 Nov 24 13:17:58 crc kubenswrapper[4824]: I1124 13:17:58.994512 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.068377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-catalog-content\") pod \"3946423b-0ffd-4991-87df-908557e14ff7\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.068446 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-utilities\") pod \"3946423b-0ffd-4991-87df-908557e14ff7\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.068493 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx4v9\" (UniqueName: \"kubernetes.io/projected/3946423b-0ffd-4991-87df-908557e14ff7-kube-api-access-zx4v9\") pod \"3946423b-0ffd-4991-87df-908557e14ff7\" (UID: \"3946423b-0ffd-4991-87df-908557e14ff7\") " Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.069930 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-utilities" (OuterVolumeSpecName: "utilities") pod "3946423b-0ffd-4991-87df-908557e14ff7" (UID: "3946423b-0ffd-4991-87df-908557e14ff7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.068796 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpk2s" event={"ID":"51db319d-7582-4e5f-8b7e-b0f11a9e7757","Type":"ContainerStarted","Data":"8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5"} Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.080685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfshk" event={"ID":"a0b31fef-5fc1-4783-ba71-a02a9b01b21a","Type":"ContainerStarted","Data":"7eb66a30e2aa15dd6942a1cc5f4a400d4965981588c8e7d1379266c91fa68a0d"} Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.083626 4824 generic.go:334] "Generic (PLEG): container finished" podID="3946423b-0ffd-4991-87df-908557e14ff7" containerID="6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d" exitCode=0 Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.083678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtmpf" event={"ID":"3946423b-0ffd-4991-87df-908557e14ff7","Type":"ContainerDied","Data":"6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d"} Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.083700 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtmpf" event={"ID":"3946423b-0ffd-4991-87df-908557e14ff7","Type":"ContainerDied","Data":"6f6922125f7f826dc5dbc1f05d43589551832c53e3ebefa13f4674d8092f6d78"} Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.083715 4824 scope.go:117] "RemoveContainer" containerID="6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.083823 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtmpf" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.088591 4824 generic.go:334] "Generic (PLEG): container finished" podID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerID="b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7" exitCode=0 Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.088641 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwsjh" event={"ID":"54a682b5-c0e0-4432-b69e-a98f0f94bd0c","Type":"ContainerDied","Data":"b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7"} Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.094082 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3946423b-0ffd-4991-87df-908557e14ff7-kube-api-access-zx4v9" (OuterVolumeSpecName: "kube-api-access-zx4v9") pod "3946423b-0ffd-4991-87df-908557e14ff7" (UID: "3946423b-0ffd-4991-87df-908557e14ff7"). InnerVolumeSpecName "kube-api-access-zx4v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.100979 4824 generic.go:334] "Generic (PLEG): container finished" podID="b8106bbc-8aca-4500-8ca9-38766f039302" containerID="ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66" exitCode=0 Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.101307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hvw9" event={"ID":"b8106bbc-8aca-4500-8ca9-38766f039302","Type":"ContainerDied","Data":"ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66"} Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.101343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hvw9" event={"ID":"b8106bbc-8aca-4500-8ca9-38766f039302","Type":"ContainerDied","Data":"49cb0547efd2902a4caa4e4c5d9165647c3752834e128b188b0e073fcc45f8bb"} Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.101414 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hvw9" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.113087 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zpk2s" podStartSLOduration=2.771759889 podStartE2EDuration="56.113071319s" podCreationTimestamp="2025-11-24 13:17:03 +0000 UTC" firstStartedPulling="2025-11-24 13:17:05.129691767 +0000 UTC m=+166.769231077" lastFinishedPulling="2025-11-24 13:17:58.471003197 +0000 UTC m=+220.110542507" observedRunningTime="2025-11-24 13:17:59.094617596 +0000 UTC m=+220.734156906" watchObservedRunningTime="2025-11-24 13:17:59.113071319 +0000 UTC m=+220.752610629" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.127444 4824 scope.go:117] "RemoveContainer" containerID="743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.155109 4824 scope.go:117] "RemoveContainer" containerID="7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.170500 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.170531 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx4v9\" (UniqueName: \"kubernetes.io/projected/3946423b-0ffd-4991-87df-908557e14ff7-kube-api-access-zx4v9\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.175005 4824 scope.go:117] "RemoveContainer" containerID="6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.175912 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hvw9"] Nov 24 13:17:59 crc kubenswrapper[4824]: E1124 13:17:59.175935 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d\": container with ID starting with 6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d not found: ID does not exist" containerID="6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.175970 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d"} err="failed to get container status \"6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d\": rpc error: code = NotFound desc = could not find container \"6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d\": container with ID starting with 6123c05c44b5ad2adede7c0f64d53cb2309b0054540bc349cebb3de993ae544d not found: ID does not exist" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.176014 4824 scope.go:117] "RemoveContainer" containerID="743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368" Nov 24 13:17:59 crc kubenswrapper[4824]: E1124 13:17:59.176664 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368\": container with ID starting with 743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368 not found: ID does not exist" containerID="743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.176695 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368"} err="failed to get container status \"743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368\": rpc error: code = NotFound desc = could not find container \"743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368\": container with ID starting with 743492baefeb3fdf46d84338ccefd617c18d065a83b263f61b270f916f14a368 not found: ID does not exist" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.176721 4824 scope.go:117] "RemoveContainer" containerID="7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246" Nov 24 13:17:59 crc kubenswrapper[4824]: E1124 13:17:59.177129 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246\": container with ID starting with 7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246 not found: ID does not exist" containerID="7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.177158 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246"} err="failed to get container status \"7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246\": rpc error: code = NotFound desc = could not find container \"7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246\": container with ID starting with 7242502b42fe671fb76e06b9ebc43dc4c85e1ac7a18bd77180fa1c51836ae246 not found: ID does not exist" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.177177 4824 scope.go:117] "RemoveContainer" containerID="ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.198867 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hvw9"] Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.200038 4824 scope.go:117] "RemoveContainer" containerID="6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.218216 4824 scope.go:117] "RemoveContainer" containerID="30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.237533 4824 scope.go:117] "RemoveContainer" containerID="ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66" Nov 24 13:17:59 crc kubenswrapper[4824]: E1124 13:17:59.239329 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66\": container with ID starting with ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66 not found: ID does not exist" containerID="ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.239378 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66"} err="failed to get container status \"ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66\": rpc error: code = NotFound desc = could not find container \"ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66\": container with ID starting with ee4085fbcf58749921a7a1f2747ca7705d2560983f5ab2bbecd0593732c7cf66 not found: ID does not exist" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.239403 4824 scope.go:117] "RemoveContainer" containerID="6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72" Nov 24 13:17:59 crc kubenswrapper[4824]: E1124 13:17:59.243166 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72\": container with ID starting with 6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72 not found: ID does not exist" containerID="6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.243210 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72"} err="failed to get container status \"6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72\": rpc error: code = NotFound desc = could not find container \"6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72\": container with ID starting with 6b8e30815adc7caba9b307250dfdd2ef6a2dcc465ca2abf0148ae446d6d5ca72 not found: ID does not exist" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.243238 4824 scope.go:117] "RemoveContainer" containerID="30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36" Nov 24 13:17:59 crc kubenswrapper[4824]: E1124 13:17:59.243785 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36\": container with ID starting with 30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36 not found: ID does not exist" containerID="30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.243827 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36"} err="failed to get container status \"30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36\": rpc error: code = NotFound desc = could not find container \"30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36\": container with ID starting with 30cae545b9ffa882aff91dca99acf0a0b37623b16370a7611bf00b870e7a0f36 not found: ID does not exist" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.251302 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3946423b-0ffd-4991-87df-908557e14ff7" (UID: "3946423b-0ffd-4991-87df-908557e14ff7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.271676 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3946423b-0ffd-4991-87df-908557e14ff7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.414034 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vtmpf"] Nov 24 13:17:59 crc kubenswrapper[4824]: I1124 13:17:59.419202 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vtmpf"] Nov 24 13:18:00 crc kubenswrapper[4824]: I1124 13:18:00.108010 4824 generic.go:334] "Generic (PLEG): container finished" podID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerID="7eb66a30e2aa15dd6942a1cc5f4a400d4965981588c8e7d1379266c91fa68a0d" exitCode=0 Nov 24 13:18:00 crc kubenswrapper[4824]: I1124 13:18:00.108072 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfshk" event={"ID":"a0b31fef-5fc1-4783-ba71-a02a9b01b21a","Type":"ContainerDied","Data":"7eb66a30e2aa15dd6942a1cc5f4a400d4965981588c8e7d1379266c91fa68a0d"} Nov 24 13:18:01 crc kubenswrapper[4824]: I1124 13:18:01.015906 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3946423b-0ffd-4991-87df-908557e14ff7" path="/var/lib/kubelet/pods/3946423b-0ffd-4991-87df-908557e14ff7/volumes" Nov 24 13:18:01 crc kubenswrapper[4824]: I1124 13:18:01.016740 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8106bbc-8aca-4500-8ca9-38766f039302" path="/var/lib/kubelet/pods/b8106bbc-8aca-4500-8ca9-38766f039302/volumes" Nov 24 13:18:01 crc kubenswrapper[4824]: I1124 13:18:01.133611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwsjh" event={"ID":"54a682b5-c0e0-4432-b69e-a98f0f94bd0c","Type":"ContainerStarted","Data":"7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759"} Nov 24 13:18:01 crc kubenswrapper[4824]: I1124 13:18:01.136447 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfshk" event={"ID":"a0b31fef-5fc1-4783-ba71-a02a9b01b21a","Type":"ContainerStarted","Data":"5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6"} Nov 24 13:18:01 crc kubenswrapper[4824]: I1124 13:18:01.154246 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cwsjh" podStartSLOduration=3.043392419 podStartE2EDuration="58.154229462s" podCreationTimestamp="2025-11-24 13:17:03 +0000 UTC" firstStartedPulling="2025-11-24 13:17:05.166179829 +0000 UTC m=+166.805719139" lastFinishedPulling="2025-11-24 13:18:00.277016862 +0000 UTC m=+221.916556182" observedRunningTime="2025-11-24 13:18:01.153226937 +0000 UTC m=+222.792766247" watchObservedRunningTime="2025-11-24 13:18:01.154229462 +0000 UTC m=+222.793768772" Nov 24 13:18:01 crc kubenswrapper[4824]: I1124 13:18:01.171584 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lfshk" podStartSLOduration=3.813729362 podStartE2EDuration="59.171565487s" podCreationTimestamp="2025-11-24 13:17:02 +0000 UTC" firstStartedPulling="2025-11-24 13:17:05.135126993 +0000 UTC m=+166.774666303" lastFinishedPulling="2025-11-24 13:18:00.492963118 +0000 UTC m=+222.132502428" observedRunningTime="2025-11-24 13:18:01.170783467 +0000 UTC m=+222.810322787" watchObservedRunningTime="2025-11-24 13:18:01.171565487 +0000 UTC m=+222.811104797" Nov 24 13:18:03 crc kubenswrapper[4824]: I1124 13:18:03.233219 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:18:03 crc kubenswrapper[4824]: I1124 13:18:03.233840 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:18:03 crc kubenswrapper[4824]: I1124 13:18:03.273102 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:18:03 crc kubenswrapper[4824]: I1124 13:18:03.446839 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:18:03 crc kubenswrapper[4824]: I1124 13:18:03.446889 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:18:03 crc kubenswrapper[4824]: I1124 13:18:03.481765 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:18:03 crc kubenswrapper[4824]: I1124 13:18:03.691315 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:18:03 crc kubenswrapper[4824]: I1124 13:18:03.692246 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:18:03 crc kubenswrapper[4824]: I1124 13:18:03.731294 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:18:04 crc kubenswrapper[4824]: I1124 13:18:04.188729 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:18:04 crc kubenswrapper[4824]: I1124 13:18:04.980355 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zpk2s"] Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.160411 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zpk2s" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerName="registry-server" containerID="cri-o://8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5" gracePeriod=2 Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.510726 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.586275 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-utilities\") pod \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.586397 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-catalog-content\") pod \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.586477 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78x92\" (UniqueName: \"kubernetes.io/projected/51db319d-7582-4e5f-8b7e-b0f11a9e7757-kube-api-access-78x92\") pod \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\" (UID: \"51db319d-7582-4e5f-8b7e-b0f11a9e7757\") " Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.598296 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51db319d-7582-4e5f-8b7e-b0f11a9e7757-kube-api-access-78x92" (OuterVolumeSpecName: "kube-api-access-78x92") pod "51db319d-7582-4e5f-8b7e-b0f11a9e7757" (UID: "51db319d-7582-4e5f-8b7e-b0f11a9e7757"). InnerVolumeSpecName "kube-api-access-78x92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.602410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-utilities" (OuterVolumeSpecName: "utilities") pod "51db319d-7582-4e5f-8b7e-b0f11a9e7757" (UID: "51db319d-7582-4e5f-8b7e-b0f11a9e7757"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.634022 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51db319d-7582-4e5f-8b7e-b0f11a9e7757" (UID: "51db319d-7582-4e5f-8b7e-b0f11a9e7757"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.688275 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.688311 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78x92\" (UniqueName: \"kubernetes.io/projected/51db319d-7582-4e5f-8b7e-b0f11a9e7757-kube-api-access-78x92\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:06 crc kubenswrapper[4824]: I1124 13:18:06.688323 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51db319d-7582-4e5f-8b7e-b0f11a9e7757-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.167517 4824 generic.go:334] "Generic (PLEG): container finished" podID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerID="8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5" exitCode=0 Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.167560 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpk2s" event={"ID":"51db319d-7582-4e5f-8b7e-b0f11a9e7757","Type":"ContainerDied","Data":"8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5"} Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.167587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpk2s" event={"ID":"51db319d-7582-4e5f-8b7e-b0f11a9e7757","Type":"ContainerDied","Data":"14fc955284e01a931d2f1c1ce54336629ee633ad9f5987ccea9bf8ae594ee06d"} Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.167603 4824 scope.go:117] "RemoveContainer" containerID="8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.167610 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpk2s" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.186650 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zpk2s"] Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.188794 4824 scope.go:117] "RemoveContainer" containerID="4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.189494 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zpk2s"] Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.207029 4824 scope.go:117] "RemoveContainer" containerID="e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.223247 4824 scope.go:117] "RemoveContainer" containerID="8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5" Nov 24 13:18:07 crc kubenswrapper[4824]: E1124 13:18:07.223654 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5\": container with ID starting with 8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5 not found: ID does not exist" containerID="8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.223697 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5"} err="failed to get container status \"8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5\": rpc error: code = NotFound desc = could not find container \"8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5\": container with ID starting with 8c929cec0d23ef0ffa8f64001ac14c473dcd59c6a111616be26370e3ba2ed2e5 not found: ID does not exist" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.223724 4824 scope.go:117] "RemoveContainer" containerID="4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4" Nov 24 13:18:07 crc kubenswrapper[4824]: E1124 13:18:07.224201 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4\": container with ID starting with 4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4 not found: ID does not exist" containerID="4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.224249 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4"} err="failed to get container status \"4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4\": rpc error: code = NotFound desc = could not find container \"4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4\": container with ID starting with 4941adbcaf5f6af92ee84d7513bb2ad0b2d70ca349431a0989eadfc78d3d3df4 not found: ID does not exist" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.224268 4824 scope.go:117] "RemoveContainer" containerID="e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248" Nov 24 13:18:07 crc kubenswrapper[4824]: E1124 13:18:07.224618 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248\": container with ID starting with e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248 not found: ID does not exist" containerID="e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248" Nov 24 13:18:07 crc kubenswrapper[4824]: I1124 13:18:07.224695 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248"} err="failed to get container status \"e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248\": rpc error: code = NotFound desc = could not find container \"e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248\": container with ID starting with e2fbbb9943a5f75d8aeb33d777a8f1ccd10252cbfa5ff69885eca1755a54d248 not found: ID does not exist" Nov 24 13:18:09 crc kubenswrapper[4824]: I1124 13:18:09.016923 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" path="/var/lib/kubelet/pods/51db319d-7582-4e5f-8b7e-b0f11a9e7757/volumes" Nov 24 13:18:10 crc kubenswrapper[4824]: I1124 13:18:10.787866 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:18:10 crc kubenswrapper[4824]: I1124 13:18:10.788480 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:18:10 crc kubenswrapper[4824]: I1124 13:18:10.788603 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:18:10 crc kubenswrapper[4824]: I1124 13:18:10.789368 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:18:10 crc kubenswrapper[4824]: I1124 13:18:10.789546 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0" gracePeriod=600 Nov 24 13:18:11 crc kubenswrapper[4824]: I1124 13:18:11.197424 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0" exitCode=0 Nov 24 13:18:11 crc kubenswrapper[4824]: I1124 13:18:11.197467 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0"} Nov 24 13:18:12 crc kubenswrapper[4824]: I1124 13:18:12.205445 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"8082a30966bfe1f65bb1eba2f0cf7e370a0e6b615322720eb09d399c05053c5b"} Nov 24 13:18:12 crc kubenswrapper[4824]: I1124 13:18:12.825331 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5p99t"] Nov 24 13:18:13 crc kubenswrapper[4824]: I1124 13:18:13.271547 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:18:13 crc kubenswrapper[4824]: I1124 13:18:13.481027 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:18:13 crc kubenswrapper[4824]: I1124 13:18:13.517044 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cwsjh"] Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.222037 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cwsjh" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerName="registry-server" containerID="cri-o://7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759" gracePeriod=2 Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.566864 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.692253 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2nvn\" (UniqueName: \"kubernetes.io/projected/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-kube-api-access-c2nvn\") pod \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.692311 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-utilities\") pod \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.692379 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-catalog-content\") pod \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\" (UID: \"54a682b5-c0e0-4432-b69e-a98f0f94bd0c\") " Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.693493 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-utilities" (OuterVolumeSpecName: "utilities") pod "54a682b5-c0e0-4432-b69e-a98f0f94bd0c" (UID: "54a682b5-c0e0-4432-b69e-a98f0f94bd0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.704085 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-kube-api-access-c2nvn" (OuterVolumeSpecName: "kube-api-access-c2nvn") pod "54a682b5-c0e0-4432-b69e-a98f0f94bd0c" (UID: "54a682b5-c0e0-4432-b69e-a98f0f94bd0c"). InnerVolumeSpecName "kube-api-access-c2nvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.739520 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54a682b5-c0e0-4432-b69e-a98f0f94bd0c" (UID: "54a682b5-c0e0-4432-b69e-a98f0f94bd0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.793538 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2nvn\" (UniqueName: \"kubernetes.io/projected/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-kube-api-access-c2nvn\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.793592 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:14 crc kubenswrapper[4824]: I1124 13:18:14.793603 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a682b5-c0e0-4432-b69e-a98f0f94bd0c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.230513 4824 generic.go:334] "Generic (PLEG): container finished" podID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerID="7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759" exitCode=0 Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.230548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwsjh" event={"ID":"54a682b5-c0e0-4432-b69e-a98f0f94bd0c","Type":"ContainerDied","Data":"7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759"} Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.230607 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cwsjh" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.230632 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwsjh" event={"ID":"54a682b5-c0e0-4432-b69e-a98f0f94bd0c","Type":"ContainerDied","Data":"4ee29cbe955a54ed7b00ee0adfc872e96cbb26fa76fdbccb0115929d9e046253"} Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.230655 4824 scope.go:117] "RemoveContainer" containerID="7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.249571 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cwsjh"] Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.256857 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cwsjh"] Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.262668 4824 scope.go:117] "RemoveContainer" containerID="b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.287247 4824 scope.go:117] "RemoveContainer" containerID="9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.303562 4824 scope.go:117] "RemoveContainer" containerID="7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759" Nov 24 13:18:15 crc kubenswrapper[4824]: E1124 13:18:15.304452 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759\": container with ID starting with 7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759 not found: ID does not exist" containerID="7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.304517 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759"} err="failed to get container status \"7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759\": rpc error: code = NotFound desc = could not find container \"7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759\": container with ID starting with 7e7d5379fb15e088bd43667b7cec69d0148593a4d87d26a61048c2e2c5bb3759 not found: ID does not exist" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.304560 4824 scope.go:117] "RemoveContainer" containerID="b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7" Nov 24 13:18:15 crc kubenswrapper[4824]: E1124 13:18:15.305079 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7\": container with ID starting with b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7 not found: ID does not exist" containerID="b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.305184 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7"} err="failed to get container status \"b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7\": rpc error: code = NotFound desc = could not find container \"b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7\": container with ID starting with b21b281ba5890d658b7142701c2f838e1ee81ecb65bb48234139f03a8ccee0e7 not found: ID does not exist" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.305274 4824 scope.go:117] "RemoveContainer" containerID="9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e" Nov 24 13:18:15 crc kubenswrapper[4824]: E1124 13:18:15.305787 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e\": container with ID starting with 9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e not found: ID does not exist" containerID="9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e" Nov 24 13:18:15 crc kubenswrapper[4824]: I1124 13:18:15.305879 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e"} err="failed to get container status \"9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e\": rpc error: code = NotFound desc = could not find container \"9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e\": container with ID starting with 9f90e7d3ea4375c228f09fda988dd8d7677e66f6d773616ccc7db7ea54da6a6e not found: ID does not exist" Nov 24 13:18:17 crc kubenswrapper[4824]: I1124 13:18:17.026301 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" path="/var/lib/kubelet/pods/54a682b5-c0e0-4432-b69e-a98f0f94bd0c/volumes" Nov 24 13:18:37 crc kubenswrapper[4824]: I1124 13:18:37.848169 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" podUID="f4875b41-d0f2-48c9-a451-839cd764a9cb" containerName="oauth-openshift" containerID="cri-o://0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3" gracePeriod=15 Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.210356 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241417 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6f96647944-m2qq5"] Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241614 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8106bbc-8aca-4500-8ca9-38766f039302" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241625 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8106bbc-8aca-4500-8ca9-38766f039302" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241634 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53333b0e-04cf-4626-83a6-d7d19e93685b" containerName="pruner" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241641 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="53333b0e-04cf-4626-83a6-d7d19e93685b" containerName="pruner" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241647 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3946423b-0ffd-4991-87df-908557e14ff7" containerName="extract-content" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241654 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3946423b-0ffd-4991-87df-908557e14ff7" containerName="extract-content" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241664 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2beaa442-848d-433a-841d-a461328b0376" containerName="collect-profiles" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241669 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2beaa442-848d-433a-841d-a461328b0376" containerName="collect-profiles" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241680 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerName="extract-utilities" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241686 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerName="extract-utilities" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241697 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac20781-a074-4984-8fd5-c756344632e2" containerName="pruner" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241702 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac20781-a074-4984-8fd5-c756344632e2" containerName="pruner" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241708 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerName="extract-utilities" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241713 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerName="extract-utilities" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241722 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4875b41-d0f2-48c9-a451-839cd764a9cb" containerName="oauth-openshift" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241728 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4875b41-d0f2-48c9-a451-839cd764a9cb" containerName="oauth-openshift" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241739 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8106bbc-8aca-4500-8ca9-38766f039302" containerName="extract-utilities" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241745 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8106bbc-8aca-4500-8ca9-38766f039302" containerName="extract-utilities" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241755 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerName="extract-content" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241762 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerName="extract-content" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241773 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3946423b-0ffd-4991-87df-908557e14ff7" containerName="extract-utilities" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241778 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3946423b-0ffd-4991-87df-908557e14ff7" containerName="extract-utilities" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241785 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerName="extract-content" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241790 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerName="extract-content" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241798 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3946423b-0ffd-4991-87df-908557e14ff7" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241818 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3946423b-0ffd-4991-87df-908557e14ff7" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241826 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241831 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241838 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8106bbc-8aca-4500-8ca9-38766f039302" containerName="extract-content" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241843 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8106bbc-8aca-4500-8ca9-38766f039302" containerName="extract-content" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.241850 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241855 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241937 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3946423b-0ffd-4991-87df-908557e14ff7" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241948 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2beaa442-848d-433a-841d-a461328b0376" containerName="collect-profiles" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241957 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac20781-a074-4984-8fd5-c756344632e2" containerName="pruner" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241964 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4875b41-d0f2-48c9-a451-839cd764a9cb" containerName="oauth-openshift" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241972 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="53333b0e-04cf-4626-83a6-d7d19e93685b" containerName="pruner" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241980 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8106bbc-8aca-4500-8ca9-38766f039302" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241989 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a682b5-c0e0-4432-b69e-a98f0f94bd0c" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.241996 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="51db319d-7582-4e5f-8b7e-b0f11a9e7757" containerName="registry-server" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.242354 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.256723 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6f96647944-m2qq5"] Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281559 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281601 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kqwm\" (UniqueName: \"kubernetes.io/projected/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-kube-api-access-9kqwm\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281628 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-audit-policies\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281684 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281707 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-template-error\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281749 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-session\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281863 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-template-login\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281942 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.281964 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.282150 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.282245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-audit-dir\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.350040 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4875b41-d0f2-48c9-a451-839cd764a9cb" containerID="0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3" exitCode=0 Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.350079 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.350084 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" event={"ID":"f4875b41-d0f2-48c9-a451-839cd764a9cb","Type":"ContainerDied","Data":"0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3"} Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.350112 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5p99t" event={"ID":"f4875b41-d0f2-48c9-a451-839cd764a9cb","Type":"ContainerDied","Data":"8db400acca7af97b15189fd31ec4f87e90b9d9fa4c4015b9f5beea4647838487"} Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.350130 4824 scope.go:117] "RemoveContainer" containerID="0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.366381 4824 scope.go:117] "RemoveContainer" containerID="0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3" Nov 24 13:18:38 crc kubenswrapper[4824]: E1124 13:18:38.366977 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3\": container with ID starting with 0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3 not found: ID does not exist" containerID="0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.367037 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3"} err="failed to get container status \"0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3\": rpc error: code = NotFound desc = could not find container \"0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3\": container with ID starting with 0dc39468982f23c4d114b0b8ce7c009266ec1e9a3cfdb37a1da1e7abf12ce3a3 not found: ID does not exist" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.382599 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-serving-cert\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.382672 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-provider-selection\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.382699 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-session\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.382740 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-ocp-branding-template\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.382768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26l8s\" (UniqueName: \"kubernetes.io/projected/f4875b41-d0f2-48c9-a451-839cd764a9cb-kube-api-access-26l8s\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.382856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-login\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.382903 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-cliconfig\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.382938 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-trusted-ca-bundle\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.383693 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-router-certs\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.383728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-policies\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.383790 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-service-ca\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.383842 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-dir\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.383981 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-error\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384046 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-idp-0-file-data\") pod \"f4875b41-d0f2-48c9-a451-839cd764a9cb\" (UID: \"f4875b41-d0f2-48c9-a451-839cd764a9cb\") " Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-session\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384245 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-template-login\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384324 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384426 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-audit-dir\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384627 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kqwm\" (UniqueName: \"kubernetes.io/projected/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-kube-api-access-9kqwm\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384658 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-audit-policies\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384706 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-template-error\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.384958 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.385897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-audit-dir\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.387142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-audit-policies\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.390757 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.393384 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.394517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.394601 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.394610 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.395006 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.395692 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.395722 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.395856 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-template-error\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.396743 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.398769 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.399309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.400168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.400477 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-template-login\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.400809 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4875b41-d0f2-48c9-a451-839cd764a9cb-kube-api-access-26l8s" (OuterVolumeSpecName: "kube-api-access-26l8s") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "kube-api-access-26l8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.400919 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.401632 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.403041 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.403163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-session\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.403280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.403401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.403411 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.403709 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.403952 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f4875b41-d0f2-48c9-a451-839cd764a9cb" (UID: "f4875b41-d0f2-48c9-a451-839cd764a9cb"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.406237 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kqwm\" (UniqueName: \"kubernetes.io/projected/11f90229-6a0c-4f79-a1e8-b878b8b62fd0-kube-api-access-9kqwm\") pod \"oauth-openshift-6f96647944-m2qq5\" (UID: \"11f90229-6a0c-4f79-a1e8-b878b8b62fd0\") " pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485634 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485681 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485694 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26l8s\" (UniqueName: \"kubernetes.io/projected/f4875b41-d0f2-48c9-a451-839cd764a9cb-kube-api-access-26l8s\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485704 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485713 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485722 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485735 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485767 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485776 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485786 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4875b41-d0f2-48c9-a451-839cd764a9cb-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485795 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485804 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485826 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.485840 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f4875b41-d0f2-48c9-a451-839cd764a9cb-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.563918 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.693623 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5p99t"] Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.696123 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5p99t"] Nov 24 13:18:38 crc kubenswrapper[4824]: I1124 13:18:38.956553 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6f96647944-m2qq5"] Nov 24 13:18:39 crc kubenswrapper[4824]: I1124 13:18:39.024923 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4875b41-d0f2-48c9-a451-839cd764a9cb" path="/var/lib/kubelet/pods/f4875b41-d0f2-48c9-a451-839cd764a9cb/volumes" Nov 24 13:18:39 crc kubenswrapper[4824]: I1124 13:18:39.357673 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" event={"ID":"11f90229-6a0c-4f79-a1e8-b878b8b62fd0","Type":"ContainerStarted","Data":"bda2d4c933298e8cfbe363650c9ff7161adbb61b305122ee5a98e25e73549417"} Nov 24 13:18:39 crc kubenswrapper[4824]: I1124 13:18:39.358020 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" event={"ID":"11f90229-6a0c-4f79-a1e8-b878b8b62fd0","Type":"ContainerStarted","Data":"e65970e49c4cab8ca88c4381ff6fd9a811d4ed5255f83d60695e5d3d83823210"} Nov 24 13:18:39 crc kubenswrapper[4824]: I1124 13:18:39.358473 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:39 crc kubenswrapper[4824]: I1124 13:18:39.377458 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" podStartSLOduration=27.377426827 podStartE2EDuration="27.377426827s" podCreationTimestamp="2025-11-24 13:18:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:18:39.374615435 +0000 UTC m=+261.014154745" watchObservedRunningTime="2025-11-24 13:18:39.377426827 +0000 UTC m=+261.016966137" Nov 24 13:18:39 crc kubenswrapper[4824]: I1124 13:18:39.863795 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6f96647944-m2qq5" Nov 24 13:18:52 crc kubenswrapper[4824]: I1124 13:18:52.981272 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lfshk"] Nov 24 13:18:52 crc kubenswrapper[4824]: I1124 13:18:52.982547 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lfshk" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerName="registry-server" containerID="cri-o://5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6" gracePeriod=30 Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.062765 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nv5fz"] Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.062832 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4fpg6"] Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.063076 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" podUID="d41e0dcc-fb19-49fc-8aab-a1837193c058" containerName="marketplace-operator" containerID="cri-o://9f824d869a0bb89d9f7516011d184f9e73e98c7582daf25b82e1ec9119e1088e" gracePeriod=30 Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.065233 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nv5fz" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="registry-server" containerID="cri-o://72712b195f31bb64170f683d28255ef1b3a96edfbda5ea3d681775dd9d3a87d1" gracePeriod=30 Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.066381 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5pzc"] Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.066672 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h5pzc" podUID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerName="registry-server" containerID="cri-o://6c263c5d9cd3be467e651adba3bca87f57f2b2c3e3142647ef651bacdad15a4d" gracePeriod=30 Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.083633 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qkj9f"] Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.084538 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.089355 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p65h5"] Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.089632 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p65h5" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerName="registry-server" containerID="cri-o://67364a06df60373f7864dc0404c368c30275c6820b6630850035292c1d2817a7" gracePeriod=30 Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.099057 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qkj9f"] Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.192351 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t6d6\" (UniqueName: \"kubernetes.io/projected/7b3b4423-493d-4cc4-9533-d87011248427-kube-api-access-9t6d6\") pod \"marketplace-operator-79b997595-qkj9f\" (UID: \"7b3b4423-493d-4cc4-9533-d87011248427\") " pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.192763 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7b3b4423-493d-4cc4-9533-d87011248427-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qkj9f\" (UID: \"7b3b4423-493d-4cc4-9533-d87011248427\") " pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.192819 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b3b4423-493d-4cc4-9533-d87011248427-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qkj9f\" (UID: \"7b3b4423-493d-4cc4-9533-d87011248427\") " pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: E1124 13:18:53.235575 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6 is running failed: container process not found" containerID="5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 13:18:53 crc kubenswrapper[4824]: E1124 13:18:53.236017 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6 is running failed: container process not found" containerID="5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 13:18:53 crc kubenswrapper[4824]: E1124 13:18:53.236611 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6 is running failed: container process not found" containerID="5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 13:18:53 crc kubenswrapper[4824]: E1124 13:18:53.236653 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-lfshk" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerName="registry-server" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.293856 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b3b4423-493d-4cc4-9533-d87011248427-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qkj9f\" (UID: \"7b3b4423-493d-4cc4-9533-d87011248427\") " pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.293945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t6d6\" (UniqueName: \"kubernetes.io/projected/7b3b4423-493d-4cc4-9533-d87011248427-kube-api-access-9t6d6\") pod \"marketplace-operator-79b997595-qkj9f\" (UID: \"7b3b4423-493d-4cc4-9533-d87011248427\") " pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.293982 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7b3b4423-493d-4cc4-9533-d87011248427-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qkj9f\" (UID: \"7b3b4423-493d-4cc4-9533-d87011248427\") " pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.295328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b3b4423-493d-4cc4-9533-d87011248427-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qkj9f\" (UID: \"7b3b4423-493d-4cc4-9533-d87011248427\") " pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.299588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7b3b4423-493d-4cc4-9533-d87011248427-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qkj9f\" (UID: \"7b3b4423-493d-4cc4-9533-d87011248427\") " pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.320779 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t6d6\" (UniqueName: \"kubernetes.io/projected/7b3b4423-493d-4cc4-9533-d87011248427-kube-api-access-9t6d6\") pod \"marketplace-operator-79b997595-qkj9f\" (UID: \"7b3b4423-493d-4cc4-9533-d87011248427\") " pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.430104 4824 generic.go:334] "Generic (PLEG): container finished" podID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerID="5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6" exitCode=0 Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.430156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfshk" event={"ID":"a0b31fef-5fc1-4783-ba71-a02a9b01b21a","Type":"ContainerDied","Data":"5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6"} Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.450491 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.606147 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-nv5fz" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="registry-server" probeResult="failure" output="" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.606982 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-nv5fz" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="registry-server" probeResult="failure" output="" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.789422 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4fpg6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.44:8080/healthz\": dial tcp 10.217.0.44:8080: connect: connection refused" start-of-body= Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.789507 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" podUID="d41e0dcc-fb19-49fc-8aab-a1837193c058" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.44:8080/healthz\": dial tcp 10.217.0.44:8080: connect: connection refused" Nov 24 13:18:53 crc kubenswrapper[4824]: I1124 13:18:53.873032 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qkj9f"] Nov 24 13:18:53 crc kubenswrapper[4824]: W1124 13:18:53.881480 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b3b4423_493d_4cc4_9533_d87011248427.slice/crio-669d57f027a87e5f517b58839f82e729ff50343e4b3ad9b993067745f2fd98af WatchSource:0}: Error finding container 669d57f027a87e5f517b58839f82e729ff50343e4b3ad9b993067745f2fd98af: Status 404 returned error can't find the container with id 669d57f027a87e5f517b58839f82e729ff50343e4b3ad9b993067745f2fd98af Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.435865 4824 generic.go:334] "Generic (PLEG): container finished" podID="d41e0dcc-fb19-49fc-8aab-a1837193c058" containerID="9f824d869a0bb89d9f7516011d184f9e73e98c7582daf25b82e1ec9119e1088e" exitCode=0 Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.436009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" event={"ID":"d41e0dcc-fb19-49fc-8aab-a1837193c058","Type":"ContainerDied","Data":"9f824d869a0bb89d9f7516011d184f9e73e98c7582daf25b82e1ec9119e1088e"} Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.437587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" event={"ID":"7b3b4423-493d-4cc4-9533-d87011248427","Type":"ContainerStarted","Data":"669d57f027a87e5f517b58839f82e729ff50343e4b3ad9b993067745f2fd98af"} Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.439329 4824 generic.go:334] "Generic (PLEG): container finished" podID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerID="72712b195f31bb64170f683d28255ef1b3a96edfbda5ea3d681775dd9d3a87d1" exitCode=0 Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.439380 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv5fz" event={"ID":"384f8f7c-bd08-4d8c-bf20-8680c8d49210","Type":"ContainerDied","Data":"72712b195f31bb64170f683d28255ef1b3a96edfbda5ea3d681775dd9d3a87d1"} Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.442383 4824 generic.go:334] "Generic (PLEG): container finished" podID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerID="6c263c5d9cd3be467e651adba3bca87f57f2b2c3e3142647ef651bacdad15a4d" exitCode=0 Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.442428 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5pzc" event={"ID":"163e9060-188a-41c6-8495-1e48b84ed3ba","Type":"ContainerDied","Data":"6c263c5d9cd3be467e651adba3bca87f57f2b2c3e3142647ef651bacdad15a4d"} Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.444074 4824 generic.go:334] "Generic (PLEG): container finished" podID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerID="67364a06df60373f7864dc0404c368c30275c6820b6630850035292c1d2817a7" exitCode=0 Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.444095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p65h5" event={"ID":"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf","Type":"ContainerDied","Data":"67364a06df60373f7864dc0404c368c30275c6820b6630850035292c1d2817a7"} Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.868831 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.933412 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2d9n\" (UniqueName: \"kubernetes.io/projected/d41e0dcc-fb19-49fc-8aab-a1837193c058-kube-api-access-h2d9n\") pod \"d41e0dcc-fb19-49fc-8aab-a1837193c058\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.933566 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-trusted-ca\") pod \"d41e0dcc-fb19-49fc-8aab-a1837193c058\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.934652 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "d41e0dcc-fb19-49fc-8aab-a1837193c058" (UID: "d41e0dcc-fb19-49fc-8aab-a1837193c058"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.937226 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-operator-metrics\") pod \"d41e0dcc-fb19-49fc-8aab-a1837193c058\" (UID: \"d41e0dcc-fb19-49fc-8aab-a1837193c058\") " Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.937556 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.941503 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d41e0dcc-fb19-49fc-8aab-a1837193c058-kube-api-access-h2d9n" (OuterVolumeSpecName: "kube-api-access-h2d9n") pod "d41e0dcc-fb19-49fc-8aab-a1837193c058" (UID: "d41e0dcc-fb19-49fc-8aab-a1837193c058"). InnerVolumeSpecName "kube-api-access-h2d9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.943537 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "d41e0dcc-fb19-49fc-8aab-a1837193c058" (UID: "d41e0dcc-fb19-49fc-8aab-a1837193c058"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:18:54 crc kubenswrapper[4824]: I1124 13:18:54.975948 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.038482 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-catalog-content\") pod \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.038591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-utilities\") pod \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.038639 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lc8b\" (UniqueName: \"kubernetes.io/projected/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-kube-api-access-2lc8b\") pod \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\" (UID: \"a0b31fef-5fc1-4783-ba71-a02a9b01b21a\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.038858 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2d9n\" (UniqueName: \"kubernetes.io/projected/d41e0dcc-fb19-49fc-8aab-a1837193c058-kube-api-access-h2d9n\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.038875 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d41e0dcc-fb19-49fc-8aab-a1837193c058-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.041620 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-utilities" (OuterVolumeSpecName: "utilities") pod "a0b31fef-5fc1-4783-ba71-a02a9b01b21a" (UID: "a0b31fef-5fc1-4783-ba71-a02a9b01b21a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.042868 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-kube-api-access-2lc8b" (OuterVolumeSpecName: "kube-api-access-2lc8b") pod "a0b31fef-5fc1-4783-ba71-a02a9b01b21a" (UID: "a0b31fef-5fc1-4783-ba71-a02a9b01b21a"). InnerVolumeSpecName "kube-api-access-2lc8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.072734 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.087745 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.139227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-catalog-content\") pod \"163e9060-188a-41c6-8495-1e48b84ed3ba\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.139262 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4k68n\" (UniqueName: \"kubernetes.io/projected/384f8f7c-bd08-4d8c-bf20-8680c8d49210-kube-api-access-4k68n\") pod \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.139355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-utilities\") pod \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.139375 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-utilities\") pod \"163e9060-188a-41c6-8495-1e48b84ed3ba\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.139424 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-catalog-content\") pod \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\" (UID: \"384f8f7c-bd08-4d8c-bf20-8680c8d49210\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.139438 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdnql\" (UniqueName: \"kubernetes.io/projected/163e9060-188a-41c6-8495-1e48b84ed3ba-kube-api-access-wdnql\") pod \"163e9060-188a-41c6-8495-1e48b84ed3ba\" (UID: \"163e9060-188a-41c6-8495-1e48b84ed3ba\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.139608 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.139621 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lc8b\" (UniqueName: \"kubernetes.io/projected/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-kube-api-access-2lc8b\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.140673 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-utilities" (OuterVolumeSpecName: "utilities") pod "384f8f7c-bd08-4d8c-bf20-8680c8d49210" (UID: "384f8f7c-bd08-4d8c-bf20-8680c8d49210"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.143495 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-utilities" (OuterVolumeSpecName: "utilities") pod "163e9060-188a-41c6-8495-1e48b84ed3ba" (UID: "163e9060-188a-41c6-8495-1e48b84ed3ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.144604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/163e9060-188a-41c6-8495-1e48b84ed3ba-kube-api-access-wdnql" (OuterVolumeSpecName: "kube-api-access-wdnql") pod "163e9060-188a-41c6-8495-1e48b84ed3ba" (UID: "163e9060-188a-41c6-8495-1e48b84ed3ba"). InnerVolumeSpecName "kube-api-access-wdnql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.145254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/384f8f7c-bd08-4d8c-bf20-8680c8d49210-kube-api-access-4k68n" (OuterVolumeSpecName: "kube-api-access-4k68n") pod "384f8f7c-bd08-4d8c-bf20-8680c8d49210" (UID: "384f8f7c-bd08-4d8c-bf20-8680c8d49210"). InnerVolumeSpecName "kube-api-access-4k68n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.171837 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "163e9060-188a-41c6-8495-1e48b84ed3ba" (UID: "163e9060-188a-41c6-8495-1e48b84ed3ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.178254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0b31fef-5fc1-4783-ba71-a02a9b01b21a" (UID: "a0b31fef-5fc1-4783-ba71-a02a9b01b21a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.224827 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "384f8f7c-bd08-4d8c-bf20-8680c8d49210" (UID: "384f8f7c-bd08-4d8c-bf20-8680c8d49210"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.240234 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdnql\" (UniqueName: \"kubernetes.io/projected/163e9060-188a-41c6-8495-1e48b84ed3ba-kube-api-access-wdnql\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.240262 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.240271 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.240279 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4k68n\" (UniqueName: \"kubernetes.io/projected/384f8f7c-bd08-4d8c-bf20-8680c8d49210-kube-api-access-4k68n\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.240288 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0b31fef-5fc1-4783-ba71-a02a9b01b21a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.240296 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/384f8f7c-bd08-4d8c-bf20-8680c8d49210-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.240304 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163e9060-188a-41c6-8495-1e48b84ed3ba-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.314582 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.443301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-catalog-content\") pod \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.443572 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z78pw\" (UniqueName: \"kubernetes.io/projected/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-kube-api-access-z78pw\") pod \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.444361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-utilities\") pod \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\" (UID: \"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf\") " Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.445255 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-utilities" (OuterVolumeSpecName: "utilities") pod "0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" (UID: "0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.445729 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.451918 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5pzc" event={"ID":"163e9060-188a-41c6-8495-1e48b84ed3ba","Type":"ContainerDied","Data":"00b10ba4175bb86230657b3eefc81621cb568b0d8d20a4a5b06eb95cea8e6a1e"} Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.451977 4824 scope.go:117] "RemoveContainer" containerID="6c263c5d9cd3be467e651adba3bca87f57f2b2c3e3142647ef651bacdad15a4d" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.452247 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5pzc" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.455287 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-kube-api-access-z78pw" (OuterVolumeSpecName: "kube-api-access-z78pw") pod "0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" (UID: "0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf"). InnerVolumeSpecName "kube-api-access-z78pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.456984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p65h5" event={"ID":"0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf","Type":"ContainerDied","Data":"681477b961afcffc08f3879139346f52204822dfa7b6391ddff8792e91188c46"} Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.457118 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p65h5" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.462137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfshk" event={"ID":"a0b31fef-5fc1-4783-ba71-a02a9b01b21a","Type":"ContainerDied","Data":"16920d1119b79b5b7c2b3b6086d0f37cfdc11eb109a9d326a42ca26f3cce16f1"} Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.462299 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lfshk" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.465599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" event={"ID":"d41e0dcc-fb19-49fc-8aab-a1837193c058","Type":"ContainerDied","Data":"078572340f12070746dca987cb02e7c4f6bc7178719db0dc2ec4c4513e7b6c1b"} Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.465621 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4fpg6" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.469356 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" event={"ID":"7b3b4423-493d-4cc4-9533-d87011248427","Type":"ContainerStarted","Data":"9e7936e49e7ad9b24e5fcab9978cf54ff337d5951c99402ef4688c4529382cbd"} Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.469648 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.476302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv5fz" event={"ID":"384f8f7c-bd08-4d8c-bf20-8680c8d49210","Type":"ContainerDied","Data":"03a681a4ad208d06aa1aedee84264adc7576ac40d06e7db0164ef43e8137198c"} Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.476436 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nv5fz" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.477272 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.489486 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qkj9f" podStartSLOduration=2.489464076 podStartE2EDuration="2.489464076s" podCreationTimestamp="2025-11-24 13:18:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:18:55.487034642 +0000 UTC m=+277.126573952" watchObservedRunningTime="2025-11-24 13:18:55.489464076 +0000 UTC m=+277.129003386" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.502070 4824 scope.go:117] "RemoveContainer" containerID="d5b61120056d86f2b996f2b97208d28f17c0d2a4bcaa1bbcdebd48342e7fa1b0" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.534613 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4fpg6"] Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.544464 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4fpg6"] Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.547234 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z78pw\" (UniqueName: \"kubernetes.io/projected/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-kube-api-access-z78pw\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.564099 4824 scope.go:117] "RemoveContainer" containerID="51cc9bb4c19d428c34ea8d09e9d18b5ec806fde1276e6b21fba036127795aebb" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.584098 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nv5fz"] Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.588260 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nv5fz"] Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.595054 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5pzc"] Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.604116 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5pzc"] Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.605354 4824 scope.go:117] "RemoveContainer" containerID="67364a06df60373f7864dc0404c368c30275c6820b6630850035292c1d2817a7" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.608519 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lfshk"] Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.611386 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lfshk"] Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.626405 4824 scope.go:117] "RemoveContainer" containerID="a1aa18757dd55cef332f0d08956c44e369f85bf46f528ae1df35e98903efc852" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.630482 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" (UID: "0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.640871 4824 scope.go:117] "RemoveContainer" containerID="d9de437acb26e2a4a56e4766b8877a424a529e5300495dae256bcaa63c2c7b25" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.648030 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.653167 4824 scope.go:117] "RemoveContainer" containerID="5d40df8edc743d863ee2bb19ddddad2cd153f40848a5f2ba90e3375176718ec6" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.664800 4824 scope.go:117] "RemoveContainer" containerID="7eb66a30e2aa15dd6942a1cc5f4a400d4965981588c8e7d1379266c91fa68a0d" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.678114 4824 scope.go:117] "RemoveContainer" containerID="54a01f87877e9239f85b8e1029def5c38fb006c38828e3a2f55114e35ed5ab8a" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.705076 4824 scope.go:117] "RemoveContainer" containerID="9f824d869a0bb89d9f7516011d184f9e73e98c7582daf25b82e1ec9119e1088e" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.716485 4824 scope.go:117] "RemoveContainer" containerID="72712b195f31bb64170f683d28255ef1b3a96edfbda5ea3d681775dd9d3a87d1" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.728667 4824 scope.go:117] "RemoveContainer" containerID="9269dc32b8598c028d9631dfc6e29e4d9b10902cde840627fe8312b17e68dc6a" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.747956 4824 scope.go:117] "RemoveContainer" containerID="d75734caa2bb85f120c9fcffb68581dd189a00ba878cf1a93d9f9eb6dc0f45cf" Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.787692 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p65h5"] Nov 24 13:18:55 crc kubenswrapper[4824]: I1124 13:18:55.790183 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p65h5"] Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.016408 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" path="/var/lib/kubelet/pods/0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf/volumes" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.017169 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="163e9060-188a-41c6-8495-1e48b84ed3ba" path="/var/lib/kubelet/pods/163e9060-188a-41c6-8495-1e48b84ed3ba/volumes" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.017869 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" path="/var/lib/kubelet/pods/384f8f7c-bd08-4d8c-bf20-8680c8d49210/volumes" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.019267 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" path="/var/lib/kubelet/pods/a0b31fef-5fc1-4783-ba71-a02a9b01b21a/volumes" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.019956 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d41e0dcc-fb19-49fc-8aab-a1837193c058" path="/var/lib/kubelet/pods/d41e0dcc-fb19-49fc-8aab-a1837193c058/volumes" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.217587 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gbfbm"] Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.217990 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="extract-content" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218030 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="extract-content" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218058 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218075 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218208 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d41e0dcc-fb19-49fc-8aab-a1837193c058" containerName="marketplace-operator" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218229 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d41e0dcc-fb19-49fc-8aab-a1837193c058" containerName="marketplace-operator" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218250 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="extract-utilities" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218266 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="extract-utilities" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218293 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218310 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218388 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerName="extract-utilities" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218409 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerName="extract-utilities" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218475 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218496 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218516 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerName="extract-content" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218534 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerName="extract-content" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218560 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerName="extract-utilities" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218577 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerName="extract-utilities" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218605 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218622 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218846 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerName="extract-utilities" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218864 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerName="extract-utilities" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218893 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerName="extract-content" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218908 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerName="extract-content" Nov 24 13:18:57 crc kubenswrapper[4824]: E1124 13:18:57.218931 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerName="extract-content" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.218945 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerName="extract-content" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.219150 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b31fef-5fc1-4783-ba71-a02a9b01b21a" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.219172 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa2b5e3-3fc7-449f-b9b1-57ee8c575edf" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.219388 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="384f8f7c-bd08-4d8c-bf20-8680c8d49210" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.219428 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="163e9060-188a-41c6-8495-1e48b84ed3ba" containerName="registry-server" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.219460 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d41e0dcc-fb19-49fc-8aab-a1837193c058" containerName="marketplace-operator" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.221428 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.225238 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.228996 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gbfbm"] Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.264763 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwlpx\" (UniqueName: \"kubernetes.io/projected/dec682c6-9c93-4e5f-a595-51a31d308c46-kube-api-access-vwlpx\") pod \"community-operators-gbfbm\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.264833 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-utilities\") pod \"community-operators-gbfbm\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.264876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-catalog-content\") pod \"community-operators-gbfbm\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.366519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-catalog-content\") pod \"community-operators-gbfbm\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.366610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwlpx\" (UniqueName: \"kubernetes.io/projected/dec682c6-9c93-4e5f-a595-51a31d308c46-kube-api-access-vwlpx\") pod \"community-operators-gbfbm\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.366638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-utilities\") pod \"community-operators-gbfbm\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.367062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-utilities\") pod \"community-operators-gbfbm\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.367154 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-catalog-content\") pod \"community-operators-gbfbm\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.387305 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwlpx\" (UniqueName: \"kubernetes.io/projected/dec682c6-9c93-4e5f-a595-51a31d308c46-kube-api-access-vwlpx\") pod \"community-operators-gbfbm\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.413533 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n8f7w"] Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.414455 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.415953 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.426769 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8f7w"] Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.468268 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-catalog-content\") pod \"redhat-operators-n8f7w\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.468320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtdtt\" (UniqueName: \"kubernetes.io/projected/ecd285b6-a57e-4499-aa95-b3bd40d67be3-kube-api-access-vtdtt\") pod \"redhat-operators-n8f7w\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.468367 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-utilities\") pod \"redhat-operators-n8f7w\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.541379 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.569704 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-catalog-content\") pod \"redhat-operators-n8f7w\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.569759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtdtt\" (UniqueName: \"kubernetes.io/projected/ecd285b6-a57e-4499-aa95-b3bd40d67be3-kube-api-access-vtdtt\") pod \"redhat-operators-n8f7w\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.569848 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-utilities\") pod \"redhat-operators-n8f7w\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.570268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-catalog-content\") pod \"redhat-operators-n8f7w\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.570559 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-utilities\") pod \"redhat-operators-n8f7w\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.592733 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtdtt\" (UniqueName: \"kubernetes.io/projected/ecd285b6-a57e-4499-aa95-b3bd40d67be3-kube-api-access-vtdtt\") pod \"redhat-operators-n8f7w\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.731548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:18:57 crc kubenswrapper[4824]: I1124 13:18:57.886429 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gbfbm"] Nov 24 13:18:57 crc kubenswrapper[4824]: W1124 13:18:57.895148 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddec682c6_9c93_4e5f_a595_51a31d308c46.slice/crio-aa518deca1a6bda27431e38794ecd38b8bc36bfd6221e7033b5f67d6afb0d946 WatchSource:0}: Error finding container aa518deca1a6bda27431e38794ecd38b8bc36bfd6221e7033b5f67d6afb0d946: Status 404 returned error can't find the container with id aa518deca1a6bda27431e38794ecd38b8bc36bfd6221e7033b5f67d6afb0d946 Nov 24 13:18:58 crc kubenswrapper[4824]: I1124 13:18:58.282890 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8f7w"] Nov 24 13:18:58 crc kubenswrapper[4824]: W1124 13:18:58.289961 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecd285b6_a57e_4499_aa95_b3bd40d67be3.slice/crio-d8ba3adc683ab4f8c7237b30272e486cfea5933c510020bffd1fcfef681770dc WatchSource:0}: Error finding container d8ba3adc683ab4f8c7237b30272e486cfea5933c510020bffd1fcfef681770dc: Status 404 returned error can't find the container with id d8ba3adc683ab4f8c7237b30272e486cfea5933c510020bffd1fcfef681770dc Nov 24 13:18:58 crc kubenswrapper[4824]: I1124 13:18:58.513063 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8f7w" event={"ID":"ecd285b6-a57e-4499-aa95-b3bd40d67be3","Type":"ContainerStarted","Data":"d8ba3adc683ab4f8c7237b30272e486cfea5933c510020bffd1fcfef681770dc"} Nov 24 13:18:58 crc kubenswrapper[4824]: I1124 13:18:58.516117 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbfbm" event={"ID":"dec682c6-9c93-4e5f-a595-51a31d308c46","Type":"ContainerStarted","Data":"aa518deca1a6bda27431e38794ecd38b8bc36bfd6221e7033b5f67d6afb0d946"} Nov 24 13:18:59 crc kubenswrapper[4824]: I1124 13:18:59.522521 4824 generic.go:334] "Generic (PLEG): container finished" podID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerID="7214804e700748f3eb18b45c62e362e34da34042d6361bcd4430216536f37997" exitCode=0 Nov 24 13:18:59 crc kubenswrapper[4824]: I1124 13:18:59.522626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbfbm" event={"ID":"dec682c6-9c93-4e5f-a595-51a31d308c46","Type":"ContainerDied","Data":"7214804e700748f3eb18b45c62e362e34da34042d6361bcd4430216536f37997"} Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.526200 4824 generic.go:334] "Generic (PLEG): container finished" podID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerID="c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395" exitCode=0 Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.526246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8f7w" event={"ID":"ecd285b6-a57e-4499-aa95-b3bd40d67be3","Type":"ContainerDied","Data":"c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395"} Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.618750 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dzj4v"] Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.619735 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.621218 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.632345 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzj4v"] Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.703687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adc5ce2-f4df-4893-9d1b-9bc4c7267faa-utilities\") pod \"redhat-marketplace-dzj4v\" (UID: \"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa\") " pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.703741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adc5ce2-f4df-4893-9d1b-9bc4c7267faa-catalog-content\") pod \"redhat-marketplace-dzj4v\" (UID: \"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa\") " pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.703778 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56kzs\" (UniqueName: \"kubernetes.io/projected/4adc5ce2-f4df-4893-9d1b-9bc4c7267faa-kube-api-access-56kzs\") pod \"redhat-marketplace-dzj4v\" (UID: \"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa\") " pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.805493 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56kzs\" (UniqueName: \"kubernetes.io/projected/4adc5ce2-f4df-4893-9d1b-9bc4c7267faa-kube-api-access-56kzs\") pod \"redhat-marketplace-dzj4v\" (UID: \"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa\") " pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.805577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adc5ce2-f4df-4893-9d1b-9bc4c7267faa-utilities\") pod \"redhat-marketplace-dzj4v\" (UID: \"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa\") " pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.805621 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adc5ce2-f4df-4893-9d1b-9bc4c7267faa-catalog-content\") pod \"redhat-marketplace-dzj4v\" (UID: \"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa\") " pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.806095 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adc5ce2-f4df-4893-9d1b-9bc4c7267faa-catalog-content\") pod \"redhat-marketplace-dzj4v\" (UID: \"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa\") " pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.806580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adc5ce2-f4df-4893-9d1b-9bc4c7267faa-utilities\") pod \"redhat-marketplace-dzj4v\" (UID: \"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa\") " pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.812517 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rqjfm"] Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.813498 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: W1124 13:18:59.815056 4824 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Nov 24 13:19:01 crc kubenswrapper[4824]: E1124 13:18:59.815085 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.827146 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rqjfm"] Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.830721 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56kzs\" (UniqueName: \"kubernetes.io/projected/4adc5ce2-f4df-4893-9d1b-9bc4c7267faa-kube-api-access-56kzs\") pod \"redhat-marketplace-dzj4v\" (UID: \"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa\") " pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.906793 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2bc66fe-f9be-4fdf-8360-62e62c629cae-utilities\") pod \"certified-operators-rqjfm\" (UID: \"e2bc66fe-f9be-4fdf-8360-62e62c629cae\") " pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.906879 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2bc66fe-f9be-4fdf-8360-62e62c629cae-catalog-content\") pod \"certified-operators-rqjfm\" (UID: \"e2bc66fe-f9be-4fdf-8360-62e62c629cae\") " pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.906924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k4tq\" (UniqueName: \"kubernetes.io/projected/e2bc66fe-f9be-4fdf-8360-62e62c629cae-kube-api-access-8k4tq\") pod \"certified-operators-rqjfm\" (UID: \"e2bc66fe-f9be-4fdf-8360-62e62c629cae\") " pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:18:59.934542 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:00.007794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k4tq\" (UniqueName: \"kubernetes.io/projected/e2bc66fe-f9be-4fdf-8360-62e62c629cae-kube-api-access-8k4tq\") pod \"certified-operators-rqjfm\" (UID: \"e2bc66fe-f9be-4fdf-8360-62e62c629cae\") " pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:00.010893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2bc66fe-f9be-4fdf-8360-62e62c629cae-utilities\") pod \"certified-operators-rqjfm\" (UID: \"e2bc66fe-f9be-4fdf-8360-62e62c629cae\") " pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:00.010962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2bc66fe-f9be-4fdf-8360-62e62c629cae-catalog-content\") pod \"certified-operators-rqjfm\" (UID: \"e2bc66fe-f9be-4fdf-8360-62e62c629cae\") " pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:00.012024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2bc66fe-f9be-4fdf-8360-62e62c629cae-catalog-content\") pod \"certified-operators-rqjfm\" (UID: \"e2bc66fe-f9be-4fdf-8360-62e62c629cae\") " pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:00.012058 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2bc66fe-f9be-4fdf-8360-62e62c629cae-utilities\") pod \"certified-operators-rqjfm\" (UID: \"e2bc66fe-f9be-4fdf-8360-62e62c629cae\") " pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:00.028385 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k4tq\" (UniqueName: \"kubernetes.io/projected/e2bc66fe-f9be-4fdf-8360-62e62c629cae-kube-api-access-8k4tq\") pod \"certified-operators-rqjfm\" (UID: \"e2bc66fe-f9be-4fdf-8360-62e62c629cae\") " pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:01.033354 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:01.034074 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:01.405175 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzj4v"] Nov 24 13:19:01 crc kubenswrapper[4824]: W1124 13:19:01.406107 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4adc5ce2_f4df_4893_9d1b_9bc4c7267faa.slice/crio-ddbc2a253f36750ba00152438e755f456248ef41b13ab86d6147e2bd01530b14 WatchSource:0}: Error finding container ddbc2a253f36750ba00152438e755f456248ef41b13ab86d6147e2bd01530b14: Status 404 returned error can't find the container with id ddbc2a253f36750ba00152438e755f456248ef41b13ab86d6147e2bd01530b14 Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:01.487300 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rqjfm"] Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:01.538907 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzj4v" event={"ID":"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa","Type":"ContainerStarted","Data":"ddbc2a253f36750ba00152438e755f456248ef41b13ab86d6147e2bd01530b14"} Nov 24 13:19:01 crc kubenswrapper[4824]: I1124 13:19:01.540001 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqjfm" event={"ID":"e2bc66fe-f9be-4fdf-8360-62e62c629cae","Type":"ContainerStarted","Data":"2006210461e5bec511aeb8c68731b9e3c1fbe8ca1d77403bec16fc95b54680c2"} Nov 24 13:19:02 crc kubenswrapper[4824]: I1124 13:19:02.546669 4824 generic.go:334] "Generic (PLEG): container finished" podID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerID="09dae5879834cb0d54d3c312d83055e60cb90aa3a5bbade3088d82fdc321578e" exitCode=0 Nov 24 13:19:02 crc kubenswrapper[4824]: I1124 13:19:02.546856 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbfbm" event={"ID":"dec682c6-9c93-4e5f-a595-51a31d308c46","Type":"ContainerDied","Data":"09dae5879834cb0d54d3c312d83055e60cb90aa3a5bbade3088d82fdc321578e"} Nov 24 13:19:02 crc kubenswrapper[4824]: I1124 13:19:02.548550 4824 generic.go:334] "Generic (PLEG): container finished" podID="4adc5ce2-f4df-4893-9d1b-9bc4c7267faa" containerID="753f33f1d1f2350390401b63118f5b5058f1ec495c2fa3be4dc31eb6d9067091" exitCode=0 Nov 24 13:19:02 crc kubenswrapper[4824]: I1124 13:19:02.548627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzj4v" event={"ID":"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa","Type":"ContainerDied","Data":"753f33f1d1f2350390401b63118f5b5058f1ec495c2fa3be4dc31eb6d9067091"} Nov 24 13:19:02 crc kubenswrapper[4824]: I1124 13:19:02.555317 4824 generic.go:334] "Generic (PLEG): container finished" podID="e2bc66fe-f9be-4fdf-8360-62e62c629cae" containerID="f279b3212fbdcfb8c748886bd8673636c9d0391fa610c4b8c8654d2290174c51" exitCode=0 Nov 24 13:19:02 crc kubenswrapper[4824]: I1124 13:19:02.555403 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqjfm" event={"ID":"e2bc66fe-f9be-4fdf-8360-62e62c629cae","Type":"ContainerDied","Data":"f279b3212fbdcfb8c748886bd8673636c9d0391fa610c4b8c8654d2290174c51"} Nov 24 13:19:02 crc kubenswrapper[4824]: I1124 13:19:02.560483 4824 generic.go:334] "Generic (PLEG): container finished" podID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerID="5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1" exitCode=0 Nov 24 13:19:02 crc kubenswrapper[4824]: I1124 13:19:02.560530 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8f7w" event={"ID":"ecd285b6-a57e-4499-aa95-b3bd40d67be3","Type":"ContainerDied","Data":"5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1"} Nov 24 13:19:03 crc kubenswrapper[4824]: I1124 13:19:03.573549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbfbm" event={"ID":"dec682c6-9c93-4e5f-a595-51a31d308c46","Type":"ContainerStarted","Data":"b0e094566c4fad9761e0c1c1f4af752c12e396e652e890e17bccf3c2c6202281"} Nov 24 13:19:03 crc kubenswrapper[4824]: I1124 13:19:03.575360 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8f7w" event={"ID":"ecd285b6-a57e-4499-aa95-b3bd40d67be3","Type":"ContainerStarted","Data":"6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8"} Nov 24 13:19:03 crc kubenswrapper[4824]: I1124 13:19:03.591927 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gbfbm" podStartSLOduration=3.112155221 podStartE2EDuration="6.591911159s" podCreationTimestamp="2025-11-24 13:18:57 +0000 UTC" firstStartedPulling="2025-11-24 13:18:59.525287101 +0000 UTC m=+281.164826431" lastFinishedPulling="2025-11-24 13:19:03.005043069 +0000 UTC m=+284.644582369" observedRunningTime="2025-11-24 13:19:03.590116612 +0000 UTC m=+285.229655922" watchObservedRunningTime="2025-11-24 13:19:03.591911159 +0000 UTC m=+285.231450469" Nov 24 13:19:03 crc kubenswrapper[4824]: I1124 13:19:03.606416 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n8f7w" podStartSLOduration=3.171848996 podStartE2EDuration="6.606398861s" podCreationTimestamp="2025-11-24 13:18:57 +0000 UTC" firstStartedPulling="2025-11-24 13:18:59.527414347 +0000 UTC m=+281.166953677" lastFinishedPulling="2025-11-24 13:19:02.961964232 +0000 UTC m=+284.601503542" observedRunningTime="2025-11-24 13:19:03.605660042 +0000 UTC m=+285.245199362" watchObservedRunningTime="2025-11-24 13:19:03.606398861 +0000 UTC m=+285.245938171" Nov 24 13:19:04 crc kubenswrapper[4824]: I1124 13:19:04.592698 4824 generic.go:334] "Generic (PLEG): container finished" podID="4adc5ce2-f4df-4893-9d1b-9bc4c7267faa" containerID="a00c5fe86a7d086aeb3d942f84c42720bac5d97a25f4f007b575b0938786084f" exitCode=0 Nov 24 13:19:04 crc kubenswrapper[4824]: I1124 13:19:04.592759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzj4v" event={"ID":"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa","Type":"ContainerDied","Data":"a00c5fe86a7d086aeb3d942f84c42720bac5d97a25f4f007b575b0938786084f"} Nov 24 13:19:04 crc kubenswrapper[4824]: I1124 13:19:04.608718 4824 generic.go:334] "Generic (PLEG): container finished" podID="e2bc66fe-f9be-4fdf-8360-62e62c629cae" containerID="4077f9fc231046c409d8d903b1720316c940d703aa62451d2a168c7f4c848c85" exitCode=0 Nov 24 13:19:04 crc kubenswrapper[4824]: I1124 13:19:04.609012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqjfm" event={"ID":"e2bc66fe-f9be-4fdf-8360-62e62c629cae","Type":"ContainerDied","Data":"4077f9fc231046c409d8d903b1720316c940d703aa62451d2a168c7f4c848c85"} Nov 24 13:19:05 crc kubenswrapper[4824]: I1124 13:19:05.615909 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzj4v" event={"ID":"4adc5ce2-f4df-4893-9d1b-9bc4c7267faa","Type":"ContainerStarted","Data":"b19beb08d4a9ce5520cc0cab383cf041f302ae0885d5437daac6733dcf23fe2c"} Nov 24 13:19:05 crc kubenswrapper[4824]: I1124 13:19:05.621121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqjfm" event={"ID":"e2bc66fe-f9be-4fdf-8360-62e62c629cae","Type":"ContainerStarted","Data":"7a6d22bf91ad66e0da605adae022bd9399d0727d88726e4220ba62f7dde59881"} Nov 24 13:19:05 crc kubenswrapper[4824]: I1124 13:19:05.640616 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dzj4v" podStartSLOduration=4.162504472 podStartE2EDuration="6.640602028s" podCreationTimestamp="2025-11-24 13:18:59 +0000 UTC" firstStartedPulling="2025-11-24 13:19:02.550729885 +0000 UTC m=+284.190269195" lastFinishedPulling="2025-11-24 13:19:05.028827441 +0000 UTC m=+286.668366751" observedRunningTime="2025-11-24 13:19:05.637001283 +0000 UTC m=+287.276540603" watchObservedRunningTime="2025-11-24 13:19:05.640602028 +0000 UTC m=+287.280141338" Nov 24 13:19:07 crc kubenswrapper[4824]: I1124 13:19:07.541619 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:19:07 crc kubenswrapper[4824]: I1124 13:19:07.541681 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:19:07 crc kubenswrapper[4824]: I1124 13:19:07.586718 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:19:07 crc kubenswrapper[4824]: I1124 13:19:07.606342 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rqjfm" podStartSLOduration=6.173734343 podStartE2EDuration="8.606324079s" podCreationTimestamp="2025-11-24 13:18:59 +0000 UTC" firstStartedPulling="2025-11-24 13:19:02.558632163 +0000 UTC m=+284.198171473" lastFinishedPulling="2025-11-24 13:19:04.991221899 +0000 UTC m=+286.630761209" observedRunningTime="2025-11-24 13:19:05.659013944 +0000 UTC m=+287.298553254" watchObservedRunningTime="2025-11-24 13:19:07.606324079 +0000 UTC m=+289.245863389" Nov 24 13:19:07 crc kubenswrapper[4824]: I1124 13:19:07.732107 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:19:07 crc kubenswrapper[4824]: I1124 13:19:07.732148 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:19:08 crc kubenswrapper[4824]: I1124 13:19:08.767829 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n8f7w" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerName="registry-server" probeResult="failure" output=< Nov 24 13:19:08 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 13:19:08 crc kubenswrapper[4824]: > Nov 24 13:19:09 crc kubenswrapper[4824]: I1124 13:19:09.935450 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:09 crc kubenswrapper[4824]: I1124 13:19:09.935540 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:09 crc kubenswrapper[4824]: I1124 13:19:09.972435 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:10 crc kubenswrapper[4824]: I1124 13:19:10.674711 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dzj4v" Nov 24 13:19:11 crc kubenswrapper[4824]: I1124 13:19:11.034718 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:11 crc kubenswrapper[4824]: I1124 13:19:11.034759 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:11 crc kubenswrapper[4824]: I1124 13:19:11.077654 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:11 crc kubenswrapper[4824]: I1124 13:19:11.687648 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rqjfm" Nov 24 13:19:17 crc kubenswrapper[4824]: I1124 13:19:17.590472 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:19:17 crc kubenswrapper[4824]: I1124 13:19:17.790990 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:19:17 crc kubenswrapper[4824]: I1124 13:19:17.847095 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:19:18 crc kubenswrapper[4824]: I1124 13:19:18.826575 4824 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 24 13:20:40 crc kubenswrapper[4824]: I1124 13:20:40.788095 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:20:40 crc kubenswrapper[4824]: I1124 13:20:40.789432 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:21:10 crc kubenswrapper[4824]: I1124 13:21:10.787700 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:21:10 crc kubenswrapper[4824]: I1124 13:21:10.788358 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:21:40 crc kubenswrapper[4824]: I1124 13:21:40.788189 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:21:40 crc kubenswrapper[4824]: I1124 13:21:40.788860 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:21:40 crc kubenswrapper[4824]: I1124 13:21:40.788916 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:21:40 crc kubenswrapper[4824]: I1124 13:21:40.789603 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8082a30966bfe1f65bb1eba2f0cf7e370a0e6b615322720eb09d399c05053c5b"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:21:40 crc kubenswrapper[4824]: I1124 13:21:40.789674 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://8082a30966bfe1f65bb1eba2f0cf7e370a0e6b615322720eb09d399c05053c5b" gracePeriod=600 Nov 24 13:21:41 crc kubenswrapper[4824]: I1124 13:21:41.278311 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="8082a30966bfe1f65bb1eba2f0cf7e370a0e6b615322720eb09d399c05053c5b" exitCode=0 Nov 24 13:21:41 crc kubenswrapper[4824]: I1124 13:21:41.278387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"8082a30966bfe1f65bb1eba2f0cf7e370a0e6b615322720eb09d399c05053c5b"} Nov 24 13:21:41 crc kubenswrapper[4824]: I1124 13:21:41.278893 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"e631cd0c8c7e50628259c7f81b9d3169967909fa981a7ac6f1bca6bbd6fcf373"} Nov 24 13:21:41 crc kubenswrapper[4824]: I1124 13:21:41.278929 4824 scope.go:117] "RemoveContainer" containerID="4a62a416fe48500ebc80dcc45996c2e93bb609a9957a0696cc5b96969f6d1de0" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.416173 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2mjns"] Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.418587 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.435325 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2mjns"] Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.535207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48e83089-0cc4-45e1-829c-1d5f2529b769-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.535280 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48e83089-0cc4-45e1-829c-1d5f2529b769-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.535420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqtfq\" (UniqueName: \"kubernetes.io/projected/48e83089-0cc4-45e1-829c-1d5f2529b769-kube-api-access-sqtfq\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.535494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48e83089-0cc4-45e1-829c-1d5f2529b769-bound-sa-token\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.535584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48e83089-0cc4-45e1-829c-1d5f2529b769-registry-tls\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.535673 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.535717 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48e83089-0cc4-45e1-829c-1d5f2529b769-registry-certificates\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.535776 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48e83089-0cc4-45e1-829c-1d5f2529b769-trusted-ca\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.557060 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.637403 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48e83089-0cc4-45e1-829c-1d5f2529b769-registry-tls\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.637481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48e83089-0cc4-45e1-829c-1d5f2529b769-registry-certificates\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.637519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48e83089-0cc4-45e1-829c-1d5f2529b769-trusted-ca\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.637562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48e83089-0cc4-45e1-829c-1d5f2529b769-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.637608 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48e83089-0cc4-45e1-829c-1d5f2529b769-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.637636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqtfq\" (UniqueName: \"kubernetes.io/projected/48e83089-0cc4-45e1-829c-1d5f2529b769-kube-api-access-sqtfq\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.637661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48e83089-0cc4-45e1-829c-1d5f2529b769-bound-sa-token\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.638239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48e83089-0cc4-45e1-829c-1d5f2529b769-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.639263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48e83089-0cc4-45e1-829c-1d5f2529b769-trusted-ca\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.639729 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48e83089-0cc4-45e1-829c-1d5f2529b769-registry-certificates\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.644938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48e83089-0cc4-45e1-829c-1d5f2529b769-registry-tls\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.645516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48e83089-0cc4-45e1-829c-1d5f2529b769-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.655748 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48e83089-0cc4-45e1-829c-1d5f2529b769-bound-sa-token\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.668766 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqtfq\" (UniqueName: \"kubernetes.io/projected/48e83089-0cc4-45e1-829c-1d5f2529b769-kube-api-access-sqtfq\") pod \"image-registry-66df7c8f76-2mjns\" (UID: \"48e83089-0cc4-45e1-829c-1d5f2529b769\") " pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.735023 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:29 crc kubenswrapper[4824]: I1124 13:22:29.907909 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2mjns"] Nov 24 13:22:30 crc kubenswrapper[4824]: I1124 13:22:30.541091 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" event={"ID":"48e83089-0cc4-45e1-829c-1d5f2529b769","Type":"ContainerStarted","Data":"6a1061c4e9c029056fa2d9c0602708e74d58e0531e60c9db64a846e51cfe1eca"} Nov 24 13:22:31 crc kubenswrapper[4824]: I1124 13:22:31.554798 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" event={"ID":"48e83089-0cc4-45e1-829c-1d5f2529b769","Type":"ContainerStarted","Data":"3f33032f5e663ec0a0781abd811c543d9b4d3b29b65098a09c5e71a98d2f1dea"} Nov 24 13:22:31 crc kubenswrapper[4824]: I1124 13:22:31.555009 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:49 crc kubenswrapper[4824]: I1124 13:22:49.741051 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" Nov 24 13:22:49 crc kubenswrapper[4824]: I1124 13:22:49.760790 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-2mjns" podStartSLOduration=20.760648989 podStartE2EDuration="20.760648989s" podCreationTimestamp="2025-11-24 13:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:22:31.579506367 +0000 UTC m=+493.219045677" watchObservedRunningTime="2025-11-24 13:22:49.760648989 +0000 UTC m=+511.400188299" Nov 24 13:22:49 crc kubenswrapper[4824]: I1124 13:22:49.782289 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zrmh8"] Nov 24 13:23:14 crc kubenswrapper[4824]: I1124 13:23:14.819596 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" podUID="b36ecad7-891d-4814-8a54-6a3c196c0525" containerName="registry" containerID="cri-o://4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf" gracePeriod=30 Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.659755 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.782631 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-certificates\") pod \"b36ecad7-891d-4814-8a54-6a3c196c0525\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.782686 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt46b\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-kube-api-access-pt46b\") pod \"b36ecad7-891d-4814-8a54-6a3c196c0525\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.782713 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-bound-sa-token\") pod \"b36ecad7-891d-4814-8a54-6a3c196c0525\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.782776 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-tls\") pod \"b36ecad7-891d-4814-8a54-6a3c196c0525\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.783109 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"b36ecad7-891d-4814-8a54-6a3c196c0525\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.783182 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-trusted-ca\") pod \"b36ecad7-891d-4814-8a54-6a3c196c0525\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.783226 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b36ecad7-891d-4814-8a54-6a3c196c0525-ca-trust-extracted\") pod \"b36ecad7-891d-4814-8a54-6a3c196c0525\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.783251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b36ecad7-891d-4814-8a54-6a3c196c0525-installation-pull-secrets\") pod \"b36ecad7-891d-4814-8a54-6a3c196c0525\" (UID: \"b36ecad7-891d-4814-8a54-6a3c196c0525\") " Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.783740 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "b36ecad7-891d-4814-8a54-6a3c196c0525" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.784677 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.784748 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "b36ecad7-891d-4814-8a54-6a3c196c0525" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.788975 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "b36ecad7-891d-4814-8a54-6a3c196c0525" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.789082 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "b36ecad7-891d-4814-8a54-6a3c196c0525" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.789229 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-kube-api-access-pt46b" (OuterVolumeSpecName: "kube-api-access-pt46b") pod "b36ecad7-891d-4814-8a54-6a3c196c0525" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525"). InnerVolumeSpecName "kube-api-access-pt46b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.789771 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36ecad7-891d-4814-8a54-6a3c196c0525-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "b36ecad7-891d-4814-8a54-6a3c196c0525" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.794926 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "b36ecad7-891d-4814-8a54-6a3c196c0525" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.795237 4824 generic.go:334] "Generic (PLEG): container finished" podID="b36ecad7-891d-4814-8a54-6a3c196c0525" containerID="4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf" exitCode=0 Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.795279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" event={"ID":"b36ecad7-891d-4814-8a54-6a3c196c0525","Type":"ContainerDied","Data":"4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf"} Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.795308 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" event={"ID":"b36ecad7-891d-4814-8a54-6a3c196c0525","Type":"ContainerDied","Data":"26e510b7ce455c9aac0a85dff7baf65416a852bbf4b8662fb4dd849bd5bdccb6"} Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.795328 4824 scope.go:117] "RemoveContainer" containerID="4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.795429 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zrmh8" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.802411 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b36ecad7-891d-4814-8a54-6a3c196c0525-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "b36ecad7-891d-4814-8a54-6a3c196c0525" (UID: "b36ecad7-891d-4814-8a54-6a3c196c0525"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.829772 4824 scope.go:117] "RemoveContainer" containerID="4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf" Nov 24 13:23:15 crc kubenswrapper[4824]: E1124 13:23:15.830351 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf\": container with ID starting with 4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf not found: ID does not exist" containerID="4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.830406 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf"} err="failed to get container status \"4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf\": rpc error: code = NotFound desc = could not find container \"4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf\": container with ID starting with 4cad5a463662155992a118c6ae8cc9c8cfe861f67e470d516188808a262048bf not found: ID does not exist" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.885740 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b36ecad7-891d-4814-8a54-6a3c196c0525-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.885777 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b36ecad7-891d-4814-8a54-6a3c196c0525-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.885794 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.885845 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt46b\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-kube-api-access-pt46b\") on node \"crc\" DevicePath \"\"" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.885861 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:23:15 crc kubenswrapper[4824]: I1124 13:23:15.885878 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b36ecad7-891d-4814-8a54-6a3c196c0525-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:23:16 crc kubenswrapper[4824]: I1124 13:23:16.124867 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zrmh8"] Nov 24 13:23:16 crc kubenswrapper[4824]: I1124 13:23:16.128957 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zrmh8"] Nov 24 13:23:17 crc kubenswrapper[4824]: I1124 13:23:17.016700 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b36ecad7-891d-4814-8a54-6a3c196c0525" path="/var/lib/kubelet/pods/b36ecad7-891d-4814-8a54-6a3c196c0525/volumes" Nov 24 13:24:10 crc kubenswrapper[4824]: I1124 13:24:10.788182 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:24:10 crc kubenswrapper[4824]: I1124 13:24:10.788769 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:24:40 crc kubenswrapper[4824]: I1124 13:24:40.788147 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:24:40 crc kubenswrapper[4824]: I1124 13:24:40.788851 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:25:10 crc kubenswrapper[4824]: I1124 13:25:10.787791 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:25:10 crc kubenswrapper[4824]: I1124 13:25:10.788578 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:25:10 crc kubenswrapper[4824]: I1124 13:25:10.788657 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:25:10 crc kubenswrapper[4824]: I1124 13:25:10.789617 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e631cd0c8c7e50628259c7f81b9d3169967909fa981a7ac6f1bca6bbd6fcf373"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:25:10 crc kubenswrapper[4824]: I1124 13:25:10.789740 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://e631cd0c8c7e50628259c7f81b9d3169967909fa981a7ac6f1bca6bbd6fcf373" gracePeriod=600 Nov 24 13:25:11 crc kubenswrapper[4824]: I1124 13:25:11.508110 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="e631cd0c8c7e50628259c7f81b9d3169967909fa981a7ac6f1bca6bbd6fcf373" exitCode=0 Nov 24 13:25:11 crc kubenswrapper[4824]: I1124 13:25:11.508192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"e631cd0c8c7e50628259c7f81b9d3169967909fa981a7ac6f1bca6bbd6fcf373"} Nov 24 13:25:11 crc kubenswrapper[4824]: I1124 13:25:11.508467 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"e56c6ac43b570680166358738b704b70ae28054c60648850ac0d5d06e92a00bf"} Nov 24 13:25:11 crc kubenswrapper[4824]: I1124 13:25:11.508489 4824 scope.go:117] "RemoveContainer" containerID="8082a30966bfe1f65bb1eba2f0cf7e370a0e6b615322720eb09d399c05053c5b" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.780311 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4hmfd"] Nov 24 13:26:16 crc kubenswrapper[4824]: E1124 13:26:16.781129 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36ecad7-891d-4814-8a54-6a3c196c0525" containerName="registry" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.781145 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36ecad7-891d-4814-8a54-6a3c196c0525" containerName="registry" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.781294 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b36ecad7-891d-4814-8a54-6a3c196c0525" containerName="registry" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.781768 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4hmfd" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.791415 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-vpkk8" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.791433 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.791705 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.798492 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4hmfd"] Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.818700 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-8mhb8"] Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.819350 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-8mhb8" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.820999 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-plcbd" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.844408 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-p8wjb"] Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.845180 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.847881 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-vl4mz" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.852549 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-8mhb8"] Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.856997 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-p8wjb"] Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.942645 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgpht\" (UniqueName: \"kubernetes.io/projected/121b83e4-cc24-4e97-831b-c05b2dce0aca-kube-api-access-tgpht\") pod \"cert-manager-5b446d88c5-8mhb8\" (UID: \"121b83e4-cc24-4e97-831b-c05b2dce0aca\") " pod="cert-manager/cert-manager-5b446d88c5-8mhb8" Nov 24 13:26:16 crc kubenswrapper[4824]: I1124 13:26:16.942739 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w6tt\" (UniqueName: \"kubernetes.io/projected/db3d2465-e3f8-434a-a279-9de75d0f2bbc-kube-api-access-7w6tt\") pod \"cert-manager-cainjector-7f985d654d-4hmfd\" (UID: \"db3d2465-e3f8-434a-a279-9de75d0f2bbc\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4hmfd" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.044073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhwmj\" (UniqueName: \"kubernetes.io/projected/c65d2285-3c4b-407b-937c-53a7d1302a8a-kube-api-access-zhwmj\") pod \"cert-manager-webhook-5655c58dd6-p8wjb\" (UID: \"c65d2285-3c4b-407b-937c-53a7d1302a8a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.044125 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgpht\" (UniqueName: \"kubernetes.io/projected/121b83e4-cc24-4e97-831b-c05b2dce0aca-kube-api-access-tgpht\") pod \"cert-manager-5b446d88c5-8mhb8\" (UID: \"121b83e4-cc24-4e97-831b-c05b2dce0aca\") " pod="cert-manager/cert-manager-5b446d88c5-8mhb8" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.044145 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w6tt\" (UniqueName: \"kubernetes.io/projected/db3d2465-e3f8-434a-a279-9de75d0f2bbc-kube-api-access-7w6tt\") pod \"cert-manager-cainjector-7f985d654d-4hmfd\" (UID: \"db3d2465-e3f8-434a-a279-9de75d0f2bbc\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4hmfd" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.062793 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgpht\" (UniqueName: \"kubernetes.io/projected/121b83e4-cc24-4e97-831b-c05b2dce0aca-kube-api-access-tgpht\") pod \"cert-manager-5b446d88c5-8mhb8\" (UID: \"121b83e4-cc24-4e97-831b-c05b2dce0aca\") " pod="cert-manager/cert-manager-5b446d88c5-8mhb8" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.070029 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w6tt\" (UniqueName: \"kubernetes.io/projected/db3d2465-e3f8-434a-a279-9de75d0f2bbc-kube-api-access-7w6tt\") pod \"cert-manager-cainjector-7f985d654d-4hmfd\" (UID: \"db3d2465-e3f8-434a-a279-9de75d0f2bbc\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4hmfd" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.102637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4hmfd" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.133820 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-8mhb8" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.145367 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhwmj\" (UniqueName: \"kubernetes.io/projected/c65d2285-3c4b-407b-937c-53a7d1302a8a-kube-api-access-zhwmj\") pod \"cert-manager-webhook-5655c58dd6-p8wjb\" (UID: \"c65d2285-3c4b-407b-937c-53a7d1302a8a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.165667 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhwmj\" (UniqueName: \"kubernetes.io/projected/c65d2285-3c4b-407b-937c-53a7d1302a8a-kube-api-access-zhwmj\") pod \"cert-manager-webhook-5655c58dd6-p8wjb\" (UID: \"c65d2285-3c4b-407b-937c-53a7d1302a8a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.167313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.347660 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4hmfd"] Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.362205 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.468045 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-p8wjb"] Nov 24 13:26:17 crc kubenswrapper[4824]: W1124 13:26:17.475488 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc65d2285_3c4b_407b_937c_53a7d1302a8a.slice/crio-467e5b791b35a2f2835513b47ae45442453f0fbd68b3a80d5fd2695d9caadc79 WatchSource:0}: Error finding container 467e5b791b35a2f2835513b47ae45442453f0fbd68b3a80d5fd2695d9caadc79: Status 404 returned error can't find the container with id 467e5b791b35a2f2835513b47ae45442453f0fbd68b3a80d5fd2695d9caadc79 Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.610216 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-8mhb8"] Nov 24 13:26:17 crc kubenswrapper[4824]: W1124 13:26:17.617496 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod121b83e4_cc24_4e97_831b_c05b2dce0aca.slice/crio-3a977be26c461a171f480cfd2ed219af14eb1b67f1116895b8658893ae14767d WatchSource:0}: Error finding container 3a977be26c461a171f480cfd2ed219af14eb1b67f1116895b8658893ae14767d: Status 404 returned error can't find the container with id 3a977be26c461a171f480cfd2ed219af14eb1b67f1116895b8658893ae14767d Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.833165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4hmfd" event={"ID":"db3d2465-e3f8-434a-a279-9de75d0f2bbc","Type":"ContainerStarted","Data":"ec9ab5034dc9c59a1397da7f3078dc937e47df75a48beef4c49441e6a967df97"} Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.834205 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" event={"ID":"c65d2285-3c4b-407b-937c-53a7d1302a8a","Type":"ContainerStarted","Data":"467e5b791b35a2f2835513b47ae45442453f0fbd68b3a80d5fd2695d9caadc79"} Nov 24 13:26:17 crc kubenswrapper[4824]: I1124 13:26:17.835059 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-8mhb8" event={"ID":"121b83e4-cc24-4e97-831b-c05b2dce0aca","Type":"ContainerStarted","Data":"3a977be26c461a171f480cfd2ed219af14eb1b67f1116895b8658893ae14767d"} Nov 24 13:26:21 crc kubenswrapper[4824]: I1124 13:26:21.857890 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" event={"ID":"c65d2285-3c4b-407b-937c-53a7d1302a8a","Type":"ContainerStarted","Data":"80de21914fae10886fb9efb628ac6ba08b355a4741a1d46eaca65adcca49a3d8"} Nov 24 13:26:21 crc kubenswrapper[4824]: I1124 13:26:21.860038 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" Nov 24 13:26:21 crc kubenswrapper[4824]: I1124 13:26:21.861115 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-8mhb8" event={"ID":"121b83e4-cc24-4e97-831b-c05b2dce0aca","Type":"ContainerStarted","Data":"0295140f1fe5162e827c562b05f7192536b17ef7f36af4936e15d1b03d360594"} Nov 24 13:26:21 crc kubenswrapper[4824]: I1124 13:26:21.862789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4hmfd" event={"ID":"db3d2465-e3f8-434a-a279-9de75d0f2bbc","Type":"ContainerStarted","Data":"178521254f48407c538baff4fe73268b072dad8df2f4728d214d216c77ee770c"} Nov 24 13:26:21 crc kubenswrapper[4824]: I1124 13:26:21.900064 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" podStartSLOduration=2.558699277 podStartE2EDuration="5.900040637s" podCreationTimestamp="2025-11-24 13:26:16 +0000 UTC" firstStartedPulling="2025-11-24 13:26:17.478824087 +0000 UTC m=+719.118363387" lastFinishedPulling="2025-11-24 13:26:20.820165427 +0000 UTC m=+722.459704747" observedRunningTime="2025-11-24 13:26:21.88473641 +0000 UTC m=+723.524275720" watchObservedRunningTime="2025-11-24 13:26:21.900040637 +0000 UTC m=+723.539579947" Nov 24 13:26:21 crc kubenswrapper[4824]: I1124 13:26:21.901578 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-4hmfd" podStartSLOduration=3.102989085 podStartE2EDuration="5.901571297s" podCreationTimestamp="2025-11-24 13:26:16 +0000 UTC" firstStartedPulling="2025-11-24 13:26:17.361988486 +0000 UTC m=+719.001527796" lastFinishedPulling="2025-11-24 13:26:20.160570698 +0000 UTC m=+721.800110008" observedRunningTime="2025-11-24 13:26:21.898955349 +0000 UTC m=+723.538494659" watchObservedRunningTime="2025-11-24 13:26:21.901571297 +0000 UTC m=+723.541110607" Nov 24 13:26:21 crc kubenswrapper[4824]: I1124 13:26:21.914562 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-8mhb8" podStartSLOduration=1.950219344 podStartE2EDuration="5.914548533s" podCreationTimestamp="2025-11-24 13:26:16 +0000 UTC" firstStartedPulling="2025-11-24 13:26:17.61964533 +0000 UTC m=+719.259184640" lastFinishedPulling="2025-11-24 13:26:21.583974519 +0000 UTC m=+723.223513829" observedRunningTime="2025-11-24 13:26:21.913033044 +0000 UTC m=+723.552572354" watchObservedRunningTime="2025-11-24 13:26:21.914548533 +0000 UTC m=+723.554087843" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.171436 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-p8wjb" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.356240 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nk5qz"] Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.356619 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovn-controller" containerID="cri-o://f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48" gracePeriod=30 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.356732 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovn-acl-logging" containerID="cri-o://ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc" gracePeriod=30 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.356716 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kube-rbac-proxy-node" containerID="cri-o://f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61" gracePeriod=30 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.356841 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff" gracePeriod=30 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.356716 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="sbdb" containerID="cri-o://a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e" gracePeriod=30 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.356877 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="nbdb" containerID="cri-o://2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60" gracePeriod=30 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.357192 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="northd" containerID="cri-o://ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92" gracePeriod=30 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.417424 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" containerID="cri-o://7e05e77c18a194c91bfe8d606ea0554e093dee70d4d5fd38ed73e5886750c7db" gracePeriod=30 Nov 24 13:26:27 crc kubenswrapper[4824]: E1124 13:26:27.706746 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f43932f_b763_4fea_9ed9_737078e6548d.slice/crio-ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f43932f_b763_4fea_9ed9_737078e6548d.slice/crio-a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e.scope\": RecentStats: unable to find data in memory cache]" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.899952 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/2.log" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.900660 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/1.log" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.900946 4824 generic.go:334] "Generic (PLEG): container finished" podID="2599a7bb-fad5-4c40-901b-189dc6687928" containerID="b79ac592a95c8705cff781ef8fb18056df813f18c8c2b01fe974b33aa5140e70" exitCode=2 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.901016 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ssstr" event={"ID":"2599a7bb-fad5-4c40-901b-189dc6687928","Type":"ContainerDied","Data":"b79ac592a95c8705cff781ef8fb18056df813f18c8c2b01fe974b33aa5140e70"} Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.901447 4824 scope.go:117] "RemoveContainer" containerID="91b514136167e06a6433e21bb9a055b91d4c479ef83a258e4d268a45a768dc0b" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.902079 4824 scope.go:117] "RemoveContainer" containerID="b79ac592a95c8705cff781ef8fb18056df813f18c8c2b01fe974b33aa5140e70" Nov 24 13:26:27 crc kubenswrapper[4824]: E1124 13:26:27.902482 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-ssstr_openshift-multus(2599a7bb-fad5-4c40-901b-189dc6687928)\"" pod="openshift-multus/multus-ssstr" podUID="2599a7bb-fad5-4c40-901b-189dc6687928" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.903662 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovnkube-controller/3.log" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.907344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovn-acl-logging/0.log" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.907920 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovn-controller/0.log" Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908363 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="7e05e77c18a194c91bfe8d606ea0554e093dee70d4d5fd38ed73e5886750c7db" exitCode=0 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908396 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e" exitCode=0 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908411 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60" exitCode=0 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908424 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92" exitCode=0 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908436 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff" exitCode=0 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"7e05e77c18a194c91bfe8d606ea0554e093dee70d4d5fd38ed73e5886750c7db"} Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908450 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61" exitCode=0 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908461 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc" exitCode=143 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908475 4824 generic.go:334] "Generic (PLEG): container finished" podID="0f43932f-b763-4fea-9ed9-737078e6548d" containerID="f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48" exitCode=143 Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908479 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e"} Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908501 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60"} Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92"} Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908526 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff"} Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908538 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61"} Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc"} Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.908561 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48"} Nov 24 13:26:27 crc kubenswrapper[4824]: I1124 13:26:27.941021 4824 scope.go:117] "RemoveContainer" containerID="377b74b0157accc2f6d9786d8004ca71bd6c9799797ded9f136379e734630b42" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.052690 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovn-acl-logging/0.log" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.053129 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovn-controller/0.log" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.053490 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105117 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pvmcv"] Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105290 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kube-rbac-proxy-node" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105303 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kube-rbac-proxy-node" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105314 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovn-acl-logging" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105321 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovn-acl-logging" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105333 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="sbdb" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105340 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="sbdb" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105351 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105358 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105370 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105379 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105388 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105393 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105402 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105408 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105415 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105421 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105429 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kubecfg-setup" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105435 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kubecfg-setup" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105446 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="nbdb" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105451 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="nbdb" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105459 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovn-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105464 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovn-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105471 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="northd" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105476 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="northd" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105589 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105600 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovn-acl-logging" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105607 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="northd" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105638 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kube-rbac-proxy-node" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105644 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105651 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105659 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="sbdb" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105665 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105673 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovn-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105680 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="nbdb" Nov 24 13:26:28 crc kubenswrapper[4824]: E1124 13:26:28.105764 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105771 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105858 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.105868 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" containerName="ovnkube-controller" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.107247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201588 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201621 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-openvswitch\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-ovn-kubernetes\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201678 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f43932f-b763-4fea-9ed9-737078e6548d-ovn-node-metrics-cert\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201705 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-etc-openvswitch\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201727 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-script-lib\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201742 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-node-log\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201761 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vgzk\" (UniqueName: \"kubernetes.io/projected/0f43932f-b763-4fea-9ed9-737078e6548d-kube-api-access-8vgzk\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-systemd\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201836 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-config\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201858 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-slash\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201873 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-netns\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201904 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-env-overrides\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-log-socket\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-bin\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201970 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-systemd-units\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.201986 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-netd\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202003 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-ovn\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202045 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-var-lib-openvswitch\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202060 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-kubelet\") pod \"0f43932f-b763-4fea-9ed9-737078e6548d\" (UID: \"0f43932f-b763-4fea-9ed9-737078e6548d\") " Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202469 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-log-socket" (OuterVolumeSpecName: "log-socket") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202505 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-node-log" (OuterVolumeSpecName: "node-log") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202651 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202690 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202725 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202771 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202793 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202838 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202856 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202877 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.202895 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.203289 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-slash" (OuterVolumeSpecName: "host-slash") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.203325 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.203432 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.203633 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.204188 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.207111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f43932f-b763-4fea-9ed9-737078e6548d-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.207765 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f43932f-b763-4fea-9ed9-737078e6548d-kube-api-access-8vgzk" (OuterVolumeSpecName: "kube-api-access-8vgzk") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "kube-api-access-8vgzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.213904 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "0f43932f-b763-4fea-9ed9-737078e6548d" (UID: "0f43932f-b763-4fea-9ed9-737078e6548d"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303478 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-run-openvswitch\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d80b0e0b-6caa-4b37-acce-af75fd4991d3-ovn-node-metrics-cert\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303573 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-kubelet\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-slash\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303619 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rbqk\" (UniqueName: \"kubernetes.io/projected/d80b0e0b-6caa-4b37-acce-af75fd4991d3-kube-api-access-9rbqk\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303642 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d80b0e0b-6caa-4b37-acce-af75fd4991d3-env-overrides\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303704 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-run-ovn\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303745 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d80b0e0b-6caa-4b37-acce-af75fd4991d3-ovnkube-script-lib\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-etc-openvswitch\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-run-ovn-kubernetes\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303824 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-run-netns\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-var-lib-openvswitch\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.303954 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-cni-netd\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304025 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-systemd-units\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-node-log\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304112 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-run-systemd\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d80b0e0b-6caa-4b37-acce-af75fd4991d3-ovnkube-config\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304169 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-log-socket\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-cni-bin\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304309 4824 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304326 4824 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304339 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304353 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f43932f-b763-4fea-9ed9-737078e6548d-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304364 4824 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304373 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304381 4824 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304389 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vgzk\" (UniqueName: \"kubernetes.io/projected/0f43932f-b763-4fea-9ed9-737078e6548d-kube-api-access-8vgzk\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304397 4824 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304405 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304413 4824 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304421 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304428 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f43932f-b763-4fea-9ed9-737078e6548d-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304437 4824 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304445 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304453 4824 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304461 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304469 4824 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304477 4824 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.304485 4824 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0f43932f-b763-4fea-9ed9-737078e6548d-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405330 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-run-systemd\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405373 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d80b0e0b-6caa-4b37-acce-af75fd4991d3-ovnkube-config\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-log-socket\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405411 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-cni-bin\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-run-openvswitch\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405450 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d80b0e0b-6caa-4b37-acce-af75fd4991d3-ovn-node-metrics-cert\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405471 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-kubelet\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405488 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-slash\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405493 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-run-systemd\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405519 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-log-socket\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rbqk\" (UniqueName: \"kubernetes.io/projected/d80b0e0b-6caa-4b37-acce-af75fd4991d3-kube-api-access-9rbqk\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d80b0e0b-6caa-4b37-acce-af75fd4991d3-env-overrides\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405625 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-run-ovn\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.406010 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.406083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-slash\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.406118 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-kubelet\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.406532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d80b0e0b-6caa-4b37-acce-af75fd4991d3-ovnkube-config\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.406569 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-cni-bin\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.405649 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.406758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d80b0e0b-6caa-4b37-acce-af75fd4991d3-env-overrides\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.408738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-run-ovn\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.408777 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-run-openvswitch\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.408822 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d80b0e0b-6caa-4b37-acce-af75fd4991d3-ovnkube-script-lib\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.408875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-etc-openvswitch\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.408918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-etc-openvswitch\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.409568 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d80b0e0b-6caa-4b37-acce-af75fd4991d3-ovnkube-script-lib\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.409615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-run-ovn-kubernetes\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.409661 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-run-ovn-kubernetes\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.409705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-run-netns\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.409735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-var-lib-openvswitch\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.409964 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-cni-netd\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.409990 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-systemd-units\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.410015 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-node-log\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.410077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-node-log\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.410106 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-run-netns\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.410132 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-var-lib-openvswitch\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.410170 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-host-cni-netd\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.410198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d80b0e0b-6caa-4b37-acce-af75fd4991d3-systemd-units\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.421297 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d80b0e0b-6caa-4b37-acce-af75fd4991d3-ovn-node-metrics-cert\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.427000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rbqk\" (UniqueName: \"kubernetes.io/projected/d80b0e0b-6caa-4b37-acce-af75fd4991d3-kube-api-access-9rbqk\") pod \"ovnkube-node-pvmcv\" (UID: \"d80b0e0b-6caa-4b37-acce-af75fd4991d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.723306 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.929341 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovn-acl-logging/0.log" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.930556 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nk5qz_0f43932f-b763-4fea-9ed9-737078e6548d/ovn-controller/0.log" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.931284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" event={"ID":"0f43932f-b763-4fea-9ed9-737078e6548d","Type":"ContainerDied","Data":"e9803f40667b9c49bc7cd16cd9f867edd39641616cbe80297f72d1cfe7e3a25c"} Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.931353 4824 scope.go:117] "RemoveContainer" containerID="7e05e77c18a194c91bfe8d606ea0554e093dee70d4d5fd38ed73e5886750c7db" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.933982 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nk5qz" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.941801 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/2.log" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.947121 4824 generic.go:334] "Generic (PLEG): container finished" podID="d80b0e0b-6caa-4b37-acce-af75fd4991d3" containerID="69cc25bea4ebf103b9cd3c1093c269e90b6fea25e05591c6b883e9d234328528" exitCode=0 Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.947185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerDied","Data":"69cc25bea4ebf103b9cd3c1093c269e90b6fea25e05591c6b883e9d234328528"} Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.947272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerStarted","Data":"50bbb6f1e5b2b97be5da270046d62fe8a8f4e4901f7034dfd95e1badd7cf8c0b"} Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.956578 4824 scope.go:117] "RemoveContainer" containerID="a54d8c35e15b317da16ce0c0b3bc7beadc8973401e78a0d0fc5ce3bad56dde0e" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.986087 4824 scope.go:117] "RemoveContainer" containerID="2b45755f799c2907f93465318fb2fc27acfca68b29864bceb1b7cd2ed4fead60" Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.990081 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nk5qz"] Nov 24 13:26:28 crc kubenswrapper[4824]: I1124 13:26:28.998283 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nk5qz"] Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.016001 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f43932f-b763-4fea-9ed9-737078e6548d" path="/var/lib/kubelet/pods/0f43932f-b763-4fea-9ed9-737078e6548d/volumes" Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.020995 4824 scope.go:117] "RemoveContainer" containerID="ba95ac4711350476b01d4c072ddd34d462d3c54459d252f96b7d6952c118fa92" Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.036379 4824 scope.go:117] "RemoveContainer" containerID="449fa53d6e9df9dbdae053bac9eff5d1dd7a2b8ef5d6b459a660b530292b4fff" Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.061282 4824 scope.go:117] "RemoveContainer" containerID="f4a5477f4692c39bded52590f615a18b557bb5f99238d7c2ff1d05d90b483c61" Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.074627 4824 scope.go:117] "RemoveContainer" containerID="ec5bc2e6f2b316b8e2ba5fff27812080f9e28b769fe3c1a3be842aff419806cc" Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.085923 4824 scope.go:117] "RemoveContainer" containerID="f967d1983d0e9671e7c8316304094d62982fff39c5a8d6b9bd37bd531fd0de48" Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.103173 4824 scope.go:117] "RemoveContainer" containerID="1fb2a3113f74b6d2078f0fe3214f8cf8b9f0caebffe4b09f911fe4094bed88d7" Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.960642 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerStarted","Data":"783706362f76920ad59b23ce725e432560ace53c0c54ebeed8e9a19a0371235d"} Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.961131 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerStarted","Data":"2175b75d2a8fbc2c3826d8619abc5d4ae1369966cc4a31c863da7bf08fb0a136"} Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.961156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerStarted","Data":"4534c6c998bf15cf7b2639deae0cc57f12be7de686f2ce0951f472d652445a20"} Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.961176 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerStarted","Data":"9d93a0a163dc90f7febf5a91fae802bc12a22ed72b84d9d5daf5620229ccb996"} Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.961198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerStarted","Data":"e33456f54a3a94860887e1b572ba37cf49f097650af6c95cdb3a8220bca5e2d9"} Nov 24 13:26:29 crc kubenswrapper[4824]: I1124 13:26:29.961217 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerStarted","Data":"b856dc0e52e0c0a1e002d10f4ed6951399ad369a959a649ac34b5bd10fd82f5c"} Nov 24 13:26:31 crc kubenswrapper[4824]: I1124 13:26:31.982150 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerStarted","Data":"8337f012f3843232a082d31c46c04b934843cfc3a836c3fb75414a70fe31b89f"} Nov 24 13:26:35 crc kubenswrapper[4824]: I1124 13:26:35.005831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" event={"ID":"d80b0e0b-6caa-4b37-acce-af75fd4991d3","Type":"ContainerStarted","Data":"98b6f46a476b935d52d1e69c1f7073ac9dea14bd33dbfd5ab4afc99d76ddf35e"} Nov 24 13:26:35 crc kubenswrapper[4824]: I1124 13:26:35.006247 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:35 crc kubenswrapper[4824]: I1124 13:26:35.006259 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:35 crc kubenswrapper[4824]: I1124 13:26:35.047202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:35 crc kubenswrapper[4824]: I1124 13:26:35.090021 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" podStartSLOduration=7.090003577 podStartE2EDuration="7.090003577s" podCreationTimestamp="2025-11-24 13:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:26:35.060285336 +0000 UTC m=+736.699824656" watchObservedRunningTime="2025-11-24 13:26:35.090003577 +0000 UTC m=+736.729542887" Nov 24 13:26:36 crc kubenswrapper[4824]: I1124 13:26:36.010755 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:36 crc kubenswrapper[4824]: I1124 13:26:36.034303 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:42 crc kubenswrapper[4824]: I1124 13:26:42.010946 4824 scope.go:117] "RemoveContainer" containerID="b79ac592a95c8705cff781ef8fb18056df813f18c8c2b01fe974b33aa5140e70" Nov 24 13:26:43 crc kubenswrapper[4824]: I1124 13:26:43.056798 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ssstr_2599a7bb-fad5-4c40-901b-189dc6687928/kube-multus/2.log" Nov 24 13:26:43 crc kubenswrapper[4824]: I1124 13:26:43.057227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ssstr" event={"ID":"2599a7bb-fad5-4c40-901b-189dc6687928","Type":"ContainerStarted","Data":"2e87a8ad4680d8b85d9f00ad1bb1d018899da8c4b8f9b32f8b0d264e4e03a105"} Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.221516 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lm6d6"] Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.222316 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" podUID="93a94b2f-1727-44d6-8438-c1492485cf06" containerName="controller-manager" containerID="cri-o://444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70" gracePeriod=30 Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.319998 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl"] Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.320228 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" podUID="831ebed0-2a1f-490a-b315-27f535e22fe0" containerName="route-controller-manager" containerID="cri-o://657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c" gracePeriod=30 Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.619446 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.669870 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.722622 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9swp\" (UniqueName: \"kubernetes.io/projected/93a94b2f-1727-44d6-8438-c1492485cf06-kube-api-access-x9swp\") pod \"93a94b2f-1727-44d6-8438-c1492485cf06\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.722713 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-config\") pod \"93a94b2f-1727-44d6-8438-c1492485cf06\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.722738 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-proxy-ca-bundles\") pod \"93a94b2f-1727-44d6-8438-c1492485cf06\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.722757 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-client-ca\") pod \"93a94b2f-1727-44d6-8438-c1492485cf06\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.722802 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93a94b2f-1727-44d6-8438-c1492485cf06-serving-cert\") pod \"93a94b2f-1727-44d6-8438-c1492485cf06\" (UID: \"93a94b2f-1727-44d6-8438-c1492485cf06\") " Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.723609 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-client-ca" (OuterVolumeSpecName: "client-ca") pod "93a94b2f-1727-44d6-8438-c1492485cf06" (UID: "93a94b2f-1727-44d6-8438-c1492485cf06"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.723861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "93a94b2f-1727-44d6-8438-c1492485cf06" (UID: "93a94b2f-1727-44d6-8438-c1492485cf06"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.723874 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-config" (OuterVolumeSpecName: "config") pod "93a94b2f-1727-44d6-8438-c1492485cf06" (UID: "93a94b2f-1727-44d6-8438-c1492485cf06"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.728612 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93a94b2f-1727-44d6-8438-c1492485cf06-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "93a94b2f-1727-44d6-8438-c1492485cf06" (UID: "93a94b2f-1727-44d6-8438-c1492485cf06"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.729276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a94b2f-1727-44d6-8438-c1492485cf06-kube-api-access-x9swp" (OuterVolumeSpecName: "kube-api-access-x9swp") pod "93a94b2f-1727-44d6-8438-c1492485cf06" (UID: "93a94b2f-1727-44d6-8438-c1492485cf06"). InnerVolumeSpecName "kube-api-access-x9swp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.824381 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-config\") pod \"831ebed0-2a1f-490a-b315-27f535e22fe0\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.825661 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/831ebed0-2a1f-490a-b315-27f535e22fe0-serving-cert\") pod \"831ebed0-2a1f-490a-b315-27f535e22fe0\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.825929 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72667\" (UniqueName: \"kubernetes.io/projected/831ebed0-2a1f-490a-b315-27f535e22fe0-kube-api-access-72667\") pod \"831ebed0-2a1f-490a-b315-27f535e22fe0\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.826156 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-client-ca\") pod \"831ebed0-2a1f-490a-b315-27f535e22fe0\" (UID: \"831ebed0-2a1f-490a-b315-27f535e22fe0\") " Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.826784 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9swp\" (UniqueName: \"kubernetes.io/projected/93a94b2f-1727-44d6-8438-c1492485cf06-kube-api-access-x9swp\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.826990 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.827256 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.827434 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93a94b2f-1727-44d6-8438-c1492485cf06-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.827629 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93a94b2f-1727-44d6-8438-c1492485cf06-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.825473 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-config" (OuterVolumeSpecName: "config") pod "831ebed0-2a1f-490a-b315-27f535e22fe0" (UID: "831ebed0-2a1f-490a-b315-27f535e22fe0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.828658 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-client-ca" (OuterVolumeSpecName: "client-ca") pod "831ebed0-2a1f-490a-b315-27f535e22fe0" (UID: "831ebed0-2a1f-490a-b315-27f535e22fe0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.832263 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/831ebed0-2a1f-490a-b315-27f535e22fe0-kube-api-access-72667" (OuterVolumeSpecName: "kube-api-access-72667") pod "831ebed0-2a1f-490a-b315-27f535e22fe0" (UID: "831ebed0-2a1f-490a-b315-27f535e22fe0"). InnerVolumeSpecName "kube-api-access-72667". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.833647 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/831ebed0-2a1f-490a-b315-27f535e22fe0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "831ebed0-2a1f-490a-b315-27f535e22fe0" (UID: "831ebed0-2a1f-490a-b315-27f535e22fe0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.929182 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.929228 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/831ebed0-2a1f-490a-b315-27f535e22fe0-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.929245 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72667\" (UniqueName: \"kubernetes.io/projected/831ebed0-2a1f-490a-b315-27f535e22fe0-kube-api-access-72667\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:57 crc kubenswrapper[4824]: I1124 13:26:57.929259 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/831ebed0-2a1f-490a-b315-27f535e22fe0-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.141331 4824 generic.go:334] "Generic (PLEG): container finished" podID="93a94b2f-1727-44d6-8438-c1492485cf06" containerID="444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70" exitCode=0 Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.141367 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" event={"ID":"93a94b2f-1727-44d6-8438-c1492485cf06","Type":"ContainerDied","Data":"444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70"} Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.141403 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.141600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lm6d6" event={"ID":"93a94b2f-1727-44d6-8438-c1492485cf06","Type":"ContainerDied","Data":"770475f891428760a1ec46278bb77e6dc845deb090e432c79f48d86402599dab"} Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.141647 4824 scope.go:117] "RemoveContainer" containerID="444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.146910 4824 generic.go:334] "Generic (PLEG): container finished" podID="831ebed0-2a1f-490a-b315-27f535e22fe0" containerID="657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c" exitCode=0 Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.147017 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.146981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" event={"ID":"831ebed0-2a1f-490a-b315-27f535e22fe0","Type":"ContainerDied","Data":"657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c"} Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.147260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl" event={"ID":"831ebed0-2a1f-490a-b315-27f535e22fe0","Type":"ContainerDied","Data":"bf2e7794b8481bb005a8ca8651f7b55becf31fc6fca4729d53fd3cf908bde62c"} Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.179854 4824 scope.go:117] "RemoveContainer" containerID="444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.180513 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lm6d6"] Nov 24 13:26:58 crc kubenswrapper[4824]: E1124 13:26:58.181751 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70\": container with ID starting with 444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70 not found: ID does not exist" containerID="444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.181963 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70"} err="failed to get container status \"444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70\": rpc error: code = NotFound desc = could not find container \"444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70\": container with ID starting with 444f3c22a8c5eeea8eb370cd1fce794a7af944f9f06f566c4471774e737b7a70 not found: ID does not exist" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.184308 4824 scope.go:117] "RemoveContainer" containerID="657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.184184 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lm6d6"] Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.195891 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl"] Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.197830 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f9ngl"] Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.203255 4824 scope.go:117] "RemoveContainer" containerID="657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c" Nov 24 13:26:58 crc kubenswrapper[4824]: E1124 13:26:58.203757 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c\": container with ID starting with 657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c not found: ID does not exist" containerID="657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.203927 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c"} err="failed to get container status \"657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c\": rpc error: code = NotFound desc = could not find container \"657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c\": container with ID starting with 657790e881bd16256164e1621a6c57b8e873f89f3129504b3a7a51167988e18c not found: ID does not exist" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.666485 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h"] Nov 24 13:26:58 crc kubenswrapper[4824]: E1124 13:26:58.666931 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="831ebed0-2a1f-490a-b315-27f535e22fe0" containerName="route-controller-manager" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.666944 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="831ebed0-2a1f-490a-b315-27f535e22fe0" containerName="route-controller-manager" Nov 24 13:26:58 crc kubenswrapper[4824]: E1124 13:26:58.666960 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a94b2f-1727-44d6-8438-c1492485cf06" containerName="controller-manager" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.666966 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a94b2f-1727-44d6-8438-c1492485cf06" containerName="controller-manager" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.667066 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="831ebed0-2a1f-490a-b315-27f535e22fe0" containerName="route-controller-manager" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.667076 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a94b2f-1727-44d6-8438-c1492485cf06" containerName="controller-manager" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.667441 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.670638 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7"] Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.671041 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.671187 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.671291 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.671442 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.671531 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.671613 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.671716 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.674411 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.674759 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.674904 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.678993 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.679035 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.679406 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.682381 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7"] Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.689326 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h"] Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.716012 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.743564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-client-ca\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.743616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-serving-cert\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.743660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzm8p\" (UniqueName: \"kubernetes.io/projected/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-kube-api-access-gzm8p\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.743685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-serving-cert\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.743706 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-config\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.743731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j96km\" (UniqueName: \"kubernetes.io/projected/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-kube-api-access-j96km\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.743750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-config\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.743770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-proxy-ca-bundles\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.743794 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-client-ca\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.767067 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pvmcv" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.844540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzm8p\" (UniqueName: \"kubernetes.io/projected/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-kube-api-access-gzm8p\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.844595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-serving-cert\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.844619 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-config\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.844643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j96km\" (UniqueName: \"kubernetes.io/projected/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-kube-api-access-j96km\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.844666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-config\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.844690 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-proxy-ca-bundles\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.844716 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-client-ca\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.844741 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-client-ca\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.844758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-serving-cert\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.846184 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-client-ca\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.846457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-proxy-ca-bundles\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.847004 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-config\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.847051 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-client-ca\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.847775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-config\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.854057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-serving-cert\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.854367 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-serving-cert\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.871544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j96km\" (UniqueName: \"kubernetes.io/projected/074220a6-c96e-4e2d-b1b6-a56e9e2c4a25-kube-api-access-j96km\") pod \"route-controller-manager-b9c4f44f-q5dt7\" (UID: \"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25\") " pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:58 crc kubenswrapper[4824]: I1124 13:26:58.888244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzm8p\" (UniqueName: \"kubernetes.io/projected/42bb0079-dfc3-4c16-916c-6d95ea37d4d8-kube-api-access-gzm8p\") pod \"controller-manager-9ffdbbff8-ldj5h\" (UID: \"42bb0079-dfc3-4c16-916c-6d95ea37d4d8\") " pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:59 crc kubenswrapper[4824]: I1124 13:26:59.010131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:26:59 crc kubenswrapper[4824]: I1124 13:26:59.016221 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="831ebed0-2a1f-490a-b315-27f535e22fe0" path="/var/lib/kubelet/pods/831ebed0-2a1f-490a-b315-27f535e22fe0/volumes" Nov 24 13:26:59 crc kubenswrapper[4824]: I1124 13:26:59.016968 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93a94b2f-1727-44d6-8438-c1492485cf06" path="/var/lib/kubelet/pods/93a94b2f-1727-44d6-8438-c1492485cf06/volumes" Nov 24 13:26:59 crc kubenswrapper[4824]: I1124 13:26:59.020632 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:26:59 crc kubenswrapper[4824]: I1124 13:26:59.475878 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7"] Nov 24 13:26:59 crc kubenswrapper[4824]: I1124 13:26:59.490045 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h"] Nov 24 13:26:59 crc kubenswrapper[4824]: W1124 13:26:59.500405 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42bb0079_dfc3_4c16_916c_6d95ea37d4d8.slice/crio-b30b3a4b6b7aa00c500940d1f0e7cc929722d6e33bc991595a28cc49f482a00d WatchSource:0}: Error finding container b30b3a4b6b7aa00c500940d1f0e7cc929722d6e33bc991595a28cc49f482a00d: Status 404 returned error can't find the container with id b30b3a4b6b7aa00c500940d1f0e7cc929722d6e33bc991595a28cc49f482a00d Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.171279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" event={"ID":"42bb0079-dfc3-4c16-916c-6d95ea37d4d8","Type":"ContainerStarted","Data":"24b4e65741f075bf7afb789e1821204af26b03deaa9c792278f5eaec24cacf7c"} Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.171628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" event={"ID":"42bb0079-dfc3-4c16-916c-6d95ea37d4d8","Type":"ContainerStarted","Data":"b30b3a4b6b7aa00c500940d1f0e7cc929722d6e33bc991595a28cc49f482a00d"} Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.171655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.172389 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" event={"ID":"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25","Type":"ContainerStarted","Data":"69e94e11d5dd385a7b773137309e32d60620c130d09cd487c7000b5b07711bdd"} Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.172407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" event={"ID":"074220a6-c96e-4e2d-b1b6-a56e9e2c4a25","Type":"ContainerStarted","Data":"ec42a4d33310911182b5a8b2946fce1d2a5af4abd64534c99b6abcaa4cc83694"} Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.173003 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.177798 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.178281 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.212840 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-b9c4f44f-q5dt7" podStartSLOduration=3.212827841 podStartE2EDuration="3.212827841s" podCreationTimestamp="2025-11-24 13:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:27:00.210885401 +0000 UTC m=+761.850424711" watchObservedRunningTime="2025-11-24 13:27:00.212827841 +0000 UTC m=+761.852367151" Nov 24 13:27:00 crc kubenswrapper[4824]: I1124 13:27:00.213065 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-9ffdbbff8-ldj5h" podStartSLOduration=3.213061557 podStartE2EDuration="3.213061557s" podCreationTimestamp="2025-11-24 13:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:27:00.197036841 +0000 UTC m=+761.836576151" watchObservedRunningTime="2025-11-24 13:27:00.213061557 +0000 UTC m=+761.852600867" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.437585 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s"] Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.439283 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.441151 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.485236 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s"] Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.547766 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67n4m\" (UniqueName: \"kubernetes.io/projected/44dcd7ab-51b6-4472-a91c-940e30d4cac9-kube-api-access-67n4m\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.548077 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.548193 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.649349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67n4m\" (UniqueName: \"kubernetes.io/projected/44dcd7ab-51b6-4472-a91c-940e30d4cac9-kube-api-access-67n4m\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.649428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.649461 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.650086 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.650326 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.667476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67n4m\" (UniqueName: \"kubernetes.io/projected/44dcd7ab-51b6-4472-a91c-940e30d4cac9-kube-api-access-67n4m\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:07 crc kubenswrapper[4824]: I1124 13:27:07.756992 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:08 crc kubenswrapper[4824]: I1124 13:27:08.179634 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s"] Nov 24 13:27:08 crc kubenswrapper[4824]: I1124 13:27:08.213875 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" event={"ID":"44dcd7ab-51b6-4472-a91c-940e30d4cac9","Type":"ContainerStarted","Data":"fd5d2821ca28ddde99e3c1c2f053bb5a2060857af109188917f0915e1a760a58"} Nov 24 13:27:09 crc kubenswrapper[4824]: I1124 13:27:09.226319 4824 generic.go:334] "Generic (PLEG): container finished" podID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerID="2ae3f61570f6c06072613ce16dc03a5cbadf1e70144cf1f1f615f2885a129edc" exitCode=0 Nov 24 13:27:09 crc kubenswrapper[4824]: I1124 13:27:09.226393 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" event={"ID":"44dcd7ab-51b6-4472-a91c-940e30d4cac9","Type":"ContainerDied","Data":"2ae3f61570f6c06072613ce16dc03a5cbadf1e70144cf1f1f615f2885a129edc"} Nov 24 13:27:09 crc kubenswrapper[4824]: I1124 13:27:09.290191 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 13:27:09 crc kubenswrapper[4824]: I1124 13:27:09.787634 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pjpp9"] Nov 24 13:27:09 crc kubenswrapper[4824]: I1124 13:27:09.788624 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:09 crc kubenswrapper[4824]: I1124 13:27:09.800418 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pjpp9"] Nov 24 13:27:09 crc kubenswrapper[4824]: I1124 13:27:09.981092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-utilities\") pod \"redhat-operators-pjpp9\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:09 crc kubenswrapper[4824]: I1124 13:27:09.981154 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-catalog-content\") pod \"redhat-operators-pjpp9\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:09 crc kubenswrapper[4824]: I1124 13:27:09.981223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7stf\" (UniqueName: \"kubernetes.io/projected/c25af346-6bf0-4e3b-91f1-d213724477a0-kube-api-access-z7stf\") pod \"redhat-operators-pjpp9\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:10 crc kubenswrapper[4824]: I1124 13:27:10.082202 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-utilities\") pod \"redhat-operators-pjpp9\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:10 crc kubenswrapper[4824]: I1124 13:27:10.082307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-catalog-content\") pod \"redhat-operators-pjpp9\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:10 crc kubenswrapper[4824]: I1124 13:27:10.082341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7stf\" (UniqueName: \"kubernetes.io/projected/c25af346-6bf0-4e3b-91f1-d213724477a0-kube-api-access-z7stf\") pod \"redhat-operators-pjpp9\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:10 crc kubenswrapper[4824]: I1124 13:27:10.082760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-utilities\") pod \"redhat-operators-pjpp9\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:10 crc kubenswrapper[4824]: I1124 13:27:10.082844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-catalog-content\") pod \"redhat-operators-pjpp9\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:10 crc kubenswrapper[4824]: I1124 13:27:10.100473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7stf\" (UniqueName: \"kubernetes.io/projected/c25af346-6bf0-4e3b-91f1-d213724477a0-kube-api-access-z7stf\") pod \"redhat-operators-pjpp9\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:10 crc kubenswrapper[4824]: I1124 13:27:10.107497 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:10 crc kubenswrapper[4824]: I1124 13:27:10.586891 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pjpp9"] Nov 24 13:27:11 crc kubenswrapper[4824]: I1124 13:27:11.240779 4824 generic.go:334] "Generic (PLEG): container finished" podID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerID="262f415fb665a4a9dcba9c9e495bff67fb4da90aec99bebb10211b420e0d3866" exitCode=0 Nov 24 13:27:11 crc kubenswrapper[4824]: I1124 13:27:11.240840 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" event={"ID":"44dcd7ab-51b6-4472-a91c-940e30d4cac9","Type":"ContainerDied","Data":"262f415fb665a4a9dcba9c9e495bff67fb4da90aec99bebb10211b420e0d3866"} Nov 24 13:27:11 crc kubenswrapper[4824]: I1124 13:27:11.242738 4824 generic.go:334] "Generic (PLEG): container finished" podID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerID="9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9" exitCode=0 Nov 24 13:27:11 crc kubenswrapper[4824]: I1124 13:27:11.242774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjpp9" event={"ID":"c25af346-6bf0-4e3b-91f1-d213724477a0","Type":"ContainerDied","Data":"9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9"} Nov 24 13:27:11 crc kubenswrapper[4824]: I1124 13:27:11.242794 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjpp9" event={"ID":"c25af346-6bf0-4e3b-91f1-d213724477a0","Type":"ContainerStarted","Data":"36ea8291abfe71038ef2895a8c5a690e1a2503aa3020cbd3af7c8e1aae7b639f"} Nov 24 13:27:12 crc kubenswrapper[4824]: I1124 13:27:12.251460 4824 generic.go:334] "Generic (PLEG): container finished" podID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerID="0a5429a0fa6e3de0fcd7c3ea48081faf981ecfa3d42b97018d394685207c69c7" exitCode=0 Nov 24 13:27:12 crc kubenswrapper[4824]: I1124 13:27:12.251496 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" event={"ID":"44dcd7ab-51b6-4472-a91c-940e30d4cac9","Type":"ContainerDied","Data":"0a5429a0fa6e3de0fcd7c3ea48081faf981ecfa3d42b97018d394685207c69c7"} Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.258649 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjpp9" event={"ID":"c25af346-6bf0-4e3b-91f1-d213724477a0","Type":"ContainerStarted","Data":"87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914"} Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.590514 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.637720 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-bundle\") pod \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.637767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-util\") pod \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.637865 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67n4m\" (UniqueName: \"kubernetes.io/projected/44dcd7ab-51b6-4472-a91c-940e30d4cac9-kube-api-access-67n4m\") pod \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\" (UID: \"44dcd7ab-51b6-4472-a91c-940e30d4cac9\") " Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.638545 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-bundle" (OuterVolumeSpecName: "bundle") pod "44dcd7ab-51b6-4472-a91c-940e30d4cac9" (UID: "44dcd7ab-51b6-4472-a91c-940e30d4cac9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.644167 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44dcd7ab-51b6-4472-a91c-940e30d4cac9-kube-api-access-67n4m" (OuterVolumeSpecName: "kube-api-access-67n4m") pod "44dcd7ab-51b6-4472-a91c-940e30d4cac9" (UID: "44dcd7ab-51b6-4472-a91c-940e30d4cac9"). InnerVolumeSpecName "kube-api-access-67n4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.651466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-util" (OuterVolumeSpecName: "util") pod "44dcd7ab-51b6-4472-a91c-940e30d4cac9" (UID: "44dcd7ab-51b6-4472-a91c-940e30d4cac9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.738879 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.738919 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44dcd7ab-51b6-4472-a91c-940e30d4cac9-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:13 crc kubenswrapper[4824]: I1124 13:27:13.738928 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67n4m\" (UniqueName: \"kubernetes.io/projected/44dcd7ab-51b6-4472-a91c-940e30d4cac9-kube-api-access-67n4m\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:14 crc kubenswrapper[4824]: I1124 13:27:14.268355 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" Nov 24 13:27:14 crc kubenswrapper[4824]: I1124 13:27:14.268950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s" event={"ID":"44dcd7ab-51b6-4472-a91c-940e30d4cac9","Type":"ContainerDied","Data":"fd5d2821ca28ddde99e3c1c2f053bb5a2060857af109188917f0915e1a760a58"} Nov 24 13:27:14 crc kubenswrapper[4824]: I1124 13:27:14.269030 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd5d2821ca28ddde99e3c1c2f053bb5a2060857af109188917f0915e1a760a58" Nov 24 13:27:14 crc kubenswrapper[4824]: I1124 13:27:14.271938 4824 generic.go:334] "Generic (PLEG): container finished" podID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerID="87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914" exitCode=0 Nov 24 13:27:14 crc kubenswrapper[4824]: I1124 13:27:14.272102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjpp9" event={"ID":"c25af346-6bf0-4e3b-91f1-d213724477a0","Type":"ContainerDied","Data":"87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914"} Nov 24 13:27:15 crc kubenswrapper[4824]: I1124 13:27:15.281003 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjpp9" event={"ID":"c25af346-6bf0-4e3b-91f1-d213724477a0","Type":"ContainerStarted","Data":"00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863"} Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.805287 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pjpp9" podStartSLOduration=6.323300646 podStartE2EDuration="9.805271163s" podCreationTimestamp="2025-11-24 13:27:09 +0000 UTC" firstStartedPulling="2025-11-24 13:27:11.245121554 +0000 UTC m=+772.884660864" lastFinishedPulling="2025-11-24 13:27:14.727092071 +0000 UTC m=+776.366631381" observedRunningTime="2025-11-24 13:27:15.317099074 +0000 UTC m=+776.956638414" watchObservedRunningTime="2025-11-24 13:27:18.805271163 +0000 UTC m=+780.444810473" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.806445 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-ghp4s"] Nov 24 13:27:18 crc kubenswrapper[4824]: E1124 13:27:18.806631 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerName="util" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.806644 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerName="util" Nov 24 13:27:18 crc kubenswrapper[4824]: E1124 13:27:18.806654 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerName="extract" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.806659 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerName="extract" Nov 24 13:27:18 crc kubenswrapper[4824]: E1124 13:27:18.806675 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerName="pull" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.806681 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerName="pull" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.806779 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="44dcd7ab-51b6-4472-a91c-940e30d4cac9" containerName="extract" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.807103 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-ghp4s" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.822021 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.822407 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.823427 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-hqbr5" Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.840011 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-ghp4s"] Nov 24 13:27:18 crc kubenswrapper[4824]: I1124 13:27:18.900024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jwkb\" (UniqueName: \"kubernetes.io/projected/04d4e05b-9e17-473f-8d25-933b9a4ed60a-kube-api-access-9jwkb\") pod \"nmstate-operator-557fdffb88-ghp4s\" (UID: \"04d4e05b-9e17-473f-8d25-933b9a4ed60a\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-ghp4s" Nov 24 13:27:19 crc kubenswrapper[4824]: I1124 13:27:19.000633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jwkb\" (UniqueName: \"kubernetes.io/projected/04d4e05b-9e17-473f-8d25-933b9a4ed60a-kube-api-access-9jwkb\") pod \"nmstate-operator-557fdffb88-ghp4s\" (UID: \"04d4e05b-9e17-473f-8d25-933b9a4ed60a\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-ghp4s" Nov 24 13:27:19 crc kubenswrapper[4824]: I1124 13:27:19.019167 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 13:27:19 crc kubenswrapper[4824]: I1124 13:27:19.029753 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 13:27:19 crc kubenswrapper[4824]: I1124 13:27:19.043905 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jwkb\" (UniqueName: \"kubernetes.io/projected/04d4e05b-9e17-473f-8d25-933b9a4ed60a-kube-api-access-9jwkb\") pod \"nmstate-operator-557fdffb88-ghp4s\" (UID: \"04d4e05b-9e17-473f-8d25-933b9a4ed60a\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-ghp4s" Nov 24 13:27:19 crc kubenswrapper[4824]: I1124 13:27:19.127782 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-hqbr5" Nov 24 13:27:19 crc kubenswrapper[4824]: I1124 13:27:19.136691 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-ghp4s" Nov 24 13:27:19 crc kubenswrapper[4824]: I1124 13:27:19.572416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-ghp4s"] Nov 24 13:27:19 crc kubenswrapper[4824]: W1124 13:27:19.581996 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04d4e05b_9e17_473f_8d25_933b9a4ed60a.slice/crio-957334a37e263fa614e05f778e2b8466ce2ca291df3c7bbf14088879a527d24d WatchSource:0}: Error finding container 957334a37e263fa614e05f778e2b8466ce2ca291df3c7bbf14088879a527d24d: Status 404 returned error can't find the container with id 957334a37e263fa614e05f778e2b8466ce2ca291df3c7bbf14088879a527d24d Nov 24 13:27:20 crc kubenswrapper[4824]: I1124 13:27:20.108615 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:20 crc kubenswrapper[4824]: I1124 13:27:20.109014 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:20 crc kubenswrapper[4824]: I1124 13:27:20.322638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-ghp4s" event={"ID":"04d4e05b-9e17-473f-8d25-933b9a4ed60a","Type":"ContainerStarted","Data":"957334a37e263fa614e05f778e2b8466ce2ca291df3c7bbf14088879a527d24d"} Nov 24 13:27:21 crc kubenswrapper[4824]: I1124 13:27:21.148902 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pjpp9" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerName="registry-server" probeResult="failure" output=< Nov 24 13:27:21 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 13:27:21 crc kubenswrapper[4824]: > Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.337379 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-ghp4s" event={"ID":"04d4e05b-9e17-473f-8d25-933b9a4ed60a","Type":"ContainerStarted","Data":"a0d4b8794838e03ed3b675a695deeedb88c1c69311ee4ba011a021b0badcd491"} Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.359964 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-ghp4s" podStartSLOduration=2.6375531800000003 podStartE2EDuration="4.359942107s" podCreationTimestamp="2025-11-24 13:27:18 +0000 UTC" firstStartedPulling="2025-11-24 13:27:19.58351482 +0000 UTC m=+781.223054130" lastFinishedPulling="2025-11-24 13:27:21.305903747 +0000 UTC m=+782.945443057" observedRunningTime="2025-11-24 13:27:22.354210058 +0000 UTC m=+783.993749408" watchObservedRunningTime="2025-11-24 13:27:22.359942107 +0000 UTC m=+783.999481457" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.587496 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jgrzs"] Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.588867 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.600609 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jgrzs"] Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.744649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2qhc\" (UniqueName: \"kubernetes.io/projected/9cab54fb-6f5d-4999-9223-d10de0017da3-kube-api-access-c2qhc\") pod \"redhat-marketplace-jgrzs\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.744696 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-utilities\") pod \"redhat-marketplace-jgrzs\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.744748 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-catalog-content\") pod \"redhat-marketplace-jgrzs\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.845499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2qhc\" (UniqueName: \"kubernetes.io/projected/9cab54fb-6f5d-4999-9223-d10de0017da3-kube-api-access-c2qhc\") pod \"redhat-marketplace-jgrzs\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.845561 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-utilities\") pod \"redhat-marketplace-jgrzs\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.845618 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-catalog-content\") pod \"redhat-marketplace-jgrzs\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.846077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-utilities\") pod \"redhat-marketplace-jgrzs\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.846166 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-catalog-content\") pod \"redhat-marketplace-jgrzs\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.865657 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2qhc\" (UniqueName: \"kubernetes.io/projected/9cab54fb-6f5d-4999-9223-d10de0017da3-kube-api-access-c2qhc\") pod \"redhat-marketplace-jgrzs\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:22 crc kubenswrapper[4824]: I1124 13:27:22.914668 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:23 crc kubenswrapper[4824]: I1124 13:27:23.377009 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jgrzs"] Nov 24 13:27:23 crc kubenswrapper[4824]: W1124 13:27:23.390154 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cab54fb_6f5d_4999_9223_d10de0017da3.slice/crio-de9186eaf4b0f245417290b024bbb9a75392552e76a4c26368037863a218aefc WatchSource:0}: Error finding container de9186eaf4b0f245417290b024bbb9a75392552e76a4c26368037863a218aefc: Status 404 returned error can't find the container with id de9186eaf4b0f245417290b024bbb9a75392552e76a4c26368037863a218aefc Nov 24 13:27:24 crc kubenswrapper[4824]: I1124 13:27:24.352708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jgrzs" event={"ID":"9cab54fb-6f5d-4999-9223-d10de0017da3","Type":"ContainerStarted","Data":"924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9"} Nov 24 13:27:24 crc kubenswrapper[4824]: I1124 13:27:24.353068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jgrzs" event={"ID":"9cab54fb-6f5d-4999-9223-d10de0017da3","Type":"ContainerStarted","Data":"de9186eaf4b0f245417290b024bbb9a75392552e76a4c26368037863a218aefc"} Nov 24 13:27:25 crc kubenswrapper[4824]: I1124 13:27:25.360284 4824 generic.go:334] "Generic (PLEG): container finished" podID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerID="924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9" exitCode=0 Nov 24 13:27:25 crc kubenswrapper[4824]: I1124 13:27:25.360354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jgrzs" event={"ID":"9cab54fb-6f5d-4999-9223-d10de0017da3","Type":"ContainerDied","Data":"924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9"} Nov 24 13:27:26 crc kubenswrapper[4824]: I1124 13:27:26.367547 4824 generic.go:334] "Generic (PLEG): container finished" podID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerID="430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f" exitCode=0 Nov 24 13:27:26 crc kubenswrapper[4824]: I1124 13:27:26.367589 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jgrzs" event={"ID":"9cab54fb-6f5d-4999-9223-d10de0017da3","Type":"ContainerDied","Data":"430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f"} Nov 24 13:27:27 crc kubenswrapper[4824]: I1124 13:27:27.375454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jgrzs" event={"ID":"9cab54fb-6f5d-4999-9223-d10de0017da3","Type":"ContainerStarted","Data":"b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83"} Nov 24 13:27:27 crc kubenswrapper[4824]: I1124 13:27:27.393890 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jgrzs" podStartSLOduration=3.9161656320000002 podStartE2EDuration="5.393869291s" podCreationTimestamp="2025-11-24 13:27:22 +0000 UTC" firstStartedPulling="2025-11-24 13:27:25.361598247 +0000 UTC m=+787.001137557" lastFinishedPulling="2025-11-24 13:27:26.839301896 +0000 UTC m=+788.478841216" observedRunningTime="2025-11-24 13:27:27.39228524 +0000 UTC m=+789.031824580" watchObservedRunningTime="2025-11-24 13:27:27.393869291 +0000 UTC m=+789.033408611" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.041099 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4"] Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.042075 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.044214 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.046266 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5"] Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.047541 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.048023 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-hw5nl" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.093877 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-gw4fj"] Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.095027 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.100448 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5"] Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.114776 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1e8318a0-5501-47b7-836a-4d650de94219-ovs-socket\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.114841 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqqm2\" (UniqueName: \"kubernetes.io/projected/1e8318a0-5501-47b7-836a-4d650de94219-kube-api-access-vqqm2\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.114875 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48zwp\" (UniqueName: \"kubernetes.io/projected/2717ebec-b84b-4d07-a1ac-67209e7e5e32-kube-api-access-48zwp\") pod \"nmstate-metrics-5dcf9c57c5-gb5z5\" (UID: \"2717ebec-b84b-4d07-a1ac-67209e7e5e32\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.114924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqlm6\" (UniqueName: \"kubernetes.io/projected/8876f622-9722-4257-8526-9f1f0e6a3daa-kube-api-access-kqlm6\") pod \"nmstate-webhook-6b89b748d8-qs6l4\" (UID: \"8876f622-9722-4257-8526-9f1f0e6a3daa\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.114953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1e8318a0-5501-47b7-836a-4d650de94219-dbus-socket\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.114989 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8876f622-9722-4257-8526-9f1f0e6a3daa-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-qs6l4\" (UID: \"8876f622-9722-4257-8526-9f1f0e6a3daa\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.115024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1e8318a0-5501-47b7-836a-4d650de94219-nmstate-lock\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.146524 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4"] Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.216740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1e8318a0-5501-47b7-836a-4d650de94219-ovs-socket\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.216798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqqm2\" (UniqueName: \"kubernetes.io/projected/1e8318a0-5501-47b7-836a-4d650de94219-kube-api-access-vqqm2\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.216837 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48zwp\" (UniqueName: \"kubernetes.io/projected/2717ebec-b84b-4d07-a1ac-67209e7e5e32-kube-api-access-48zwp\") pod \"nmstate-metrics-5dcf9c57c5-gb5z5\" (UID: \"2717ebec-b84b-4d07-a1ac-67209e7e5e32\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.216866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqlm6\" (UniqueName: \"kubernetes.io/projected/8876f622-9722-4257-8526-9f1f0e6a3daa-kube-api-access-kqlm6\") pod \"nmstate-webhook-6b89b748d8-qs6l4\" (UID: \"8876f622-9722-4257-8526-9f1f0e6a3daa\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.216887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1e8318a0-5501-47b7-836a-4d650de94219-dbus-socket\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.216904 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8876f622-9722-4257-8526-9f1f0e6a3daa-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-qs6l4\" (UID: \"8876f622-9722-4257-8526-9f1f0e6a3daa\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.216931 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1e8318a0-5501-47b7-836a-4d650de94219-nmstate-lock\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.216991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1e8318a0-5501-47b7-836a-4d650de94219-nmstate-lock\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.217024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1e8318a0-5501-47b7-836a-4d650de94219-ovs-socket\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.217694 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1e8318a0-5501-47b7-836a-4d650de94219-dbus-socket\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.223126 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8876f622-9722-4257-8526-9f1f0e6a3daa-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-qs6l4\" (UID: \"8876f622-9722-4257-8526-9f1f0e6a3daa\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.241291 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48zwp\" (UniqueName: \"kubernetes.io/projected/2717ebec-b84b-4d07-a1ac-67209e7e5e32-kube-api-access-48zwp\") pod \"nmstate-metrics-5dcf9c57c5-gb5z5\" (UID: \"2717ebec-b84b-4d07-a1ac-67209e7e5e32\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.241739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqqm2\" (UniqueName: \"kubernetes.io/projected/1e8318a0-5501-47b7-836a-4d650de94219-kube-api-access-vqqm2\") pod \"nmstate-handler-gw4fj\" (UID: \"1e8318a0-5501-47b7-836a-4d650de94219\") " pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.244227 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqlm6\" (UniqueName: \"kubernetes.io/projected/8876f622-9722-4257-8526-9f1f0e6a3daa-kube-api-access-kqlm6\") pod \"nmstate-webhook-6b89b748d8-qs6l4\" (UID: \"8876f622-9722-4257-8526-9f1f0e6a3daa\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.255409 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s"] Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.256056 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.262921 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.263121 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-cmhrj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.263244 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.274884 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s"] Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.322418 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a252dabf-79e7-442e-990c-3f15bc47d536-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.322489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4fq7\" (UniqueName: \"kubernetes.io/projected/a252dabf-79e7-442e-990c-3f15bc47d536-kube-api-access-d4fq7\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.322536 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a252dabf-79e7-442e-990c-3f15bc47d536-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.361435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.375489 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.415611 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.423070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a252dabf-79e7-442e-990c-3f15bc47d536-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.424007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a252dabf-79e7-442e-990c-3f15bc47d536-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.424121 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4fq7\" (UniqueName: \"kubernetes.io/projected/a252dabf-79e7-442e-990c-3f15bc47d536-kube-api-access-d4fq7\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.425571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a252dabf-79e7-442e-990c-3f15bc47d536-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: E1124 13:27:28.425984 4824 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 24 13:27:28 crc kubenswrapper[4824]: E1124 13:27:28.426024 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a252dabf-79e7-442e-990c-3f15bc47d536-plugin-serving-cert podName:a252dabf-79e7-442e-990c-3f15bc47d536 nodeName:}" failed. No retries permitted until 2025-11-24 13:27:28.926012434 +0000 UTC m=+790.565551734 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/a252dabf-79e7-442e-990c-3f15bc47d536-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-7sv5s" (UID: "a252dabf-79e7-442e-990c-3f15bc47d536") : secret "plugin-serving-cert" not found Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.444766 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4fq7\" (UniqueName: \"kubernetes.io/projected/a252dabf-79e7-442e-990c-3f15bc47d536-kube-api-access-d4fq7\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.494539 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6866949b48-flgn5"] Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.495455 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.515731 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6866949b48-flgn5"] Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.525431 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-console-config\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.525472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/eb457852-894e-4da1-aad8-c77c54da041b-console-oauth-config\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.525504 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb457852-894e-4da1-aad8-c77c54da041b-console-serving-cert\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.525891 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-trusted-ca-bundle\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.525933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92865\" (UniqueName: \"kubernetes.io/projected/eb457852-894e-4da1-aad8-c77c54da041b-kube-api-access-92865\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.525974 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-oauth-serving-cert\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.525998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-service-ca\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.627533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-console-config\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.627751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-console-config\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.627792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/eb457852-894e-4da1-aad8-c77c54da041b-console-oauth-config\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.628278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb457852-894e-4da1-aad8-c77c54da041b-console-serving-cert\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.628350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-trusted-ca-bundle\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.628380 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92865\" (UniqueName: \"kubernetes.io/projected/eb457852-894e-4da1-aad8-c77c54da041b-kube-api-access-92865\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.628398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-oauth-serving-cert\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.628414 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-service-ca\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.629495 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-service-ca\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.629918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-oauth-serving-cert\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.630732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb457852-894e-4da1-aad8-c77c54da041b-trusted-ca-bundle\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.637593 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb457852-894e-4da1-aad8-c77c54da041b-console-serving-cert\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.638024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/eb457852-894e-4da1-aad8-c77c54da041b-console-oauth-config\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.656286 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92865\" (UniqueName: \"kubernetes.io/projected/eb457852-894e-4da1-aad8-c77c54da041b-kube-api-access-92865\") pod \"console-6866949b48-flgn5\" (UID: \"eb457852-894e-4da1-aad8-c77c54da041b\") " pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.843202 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.932857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a252dabf-79e7-442e-990c-3f15bc47d536-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.935985 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a252dabf-79e7-442e-990c-3f15bc47d536-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-7sv5s\" (UID: \"a252dabf-79e7-442e-990c-3f15bc47d536\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:28 crc kubenswrapper[4824]: I1124 13:27:28.989073 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4"] Nov 24 13:27:29 crc kubenswrapper[4824]: W1124 13:27:29.001057 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8876f622_9722_4257_8526_9f1f0e6a3daa.slice/crio-93091bf7b5cd3984f760884b07bb5f683820facb6478ea1832d8b5a5264ae35e WatchSource:0}: Error finding container 93091bf7b5cd3984f760884b07bb5f683820facb6478ea1832d8b5a5264ae35e: Status 404 returned error can't find the container with id 93091bf7b5cd3984f760884b07bb5f683820facb6478ea1832d8b5a5264ae35e Nov 24 13:27:29 crc kubenswrapper[4824]: I1124 13:27:29.072665 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5"] Nov 24 13:27:29 crc kubenswrapper[4824]: W1124 13:27:29.078057 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2717ebec_b84b_4d07_a1ac_67209e7e5e32.slice/crio-6acfce83266a17255dcef0d70ab1f3c5c3b0cdec90d3c863c11df5e7479dbe0b WatchSource:0}: Error finding container 6acfce83266a17255dcef0d70ab1f3c5c3b0cdec90d3c863c11df5e7479dbe0b: Status 404 returned error can't find the container with id 6acfce83266a17255dcef0d70ab1f3c5c3b0cdec90d3c863c11df5e7479dbe0b Nov 24 13:27:29 crc kubenswrapper[4824]: I1124 13:27:29.196631 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" Nov 24 13:27:29 crc kubenswrapper[4824]: I1124 13:27:29.236887 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6866949b48-flgn5"] Nov 24 13:27:29 crc kubenswrapper[4824]: I1124 13:27:29.388701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5" event={"ID":"2717ebec-b84b-4d07-a1ac-67209e7e5e32","Type":"ContainerStarted","Data":"6acfce83266a17255dcef0d70ab1f3c5c3b0cdec90d3c863c11df5e7479dbe0b"} Nov 24 13:27:29 crc kubenswrapper[4824]: I1124 13:27:29.391822 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6866949b48-flgn5" event={"ID":"eb457852-894e-4da1-aad8-c77c54da041b","Type":"ContainerStarted","Data":"40b33df343fa877bc1458692b8e3118734a712996fa7e7e29badca073b9ec43f"} Nov 24 13:27:29 crc kubenswrapper[4824]: I1124 13:27:29.395022 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gw4fj" event={"ID":"1e8318a0-5501-47b7-836a-4d650de94219","Type":"ContainerStarted","Data":"f921736b78c39fb79db259be963f866ebe3a3e9db044a14111abe87960c64efd"} Nov 24 13:27:29 crc kubenswrapper[4824]: I1124 13:27:29.396010 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" event={"ID":"8876f622-9722-4257-8526-9f1f0e6a3daa","Type":"ContainerStarted","Data":"93091bf7b5cd3984f760884b07bb5f683820facb6478ea1832d8b5a5264ae35e"} Nov 24 13:27:29 crc kubenswrapper[4824]: I1124 13:27:29.639511 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s"] Nov 24 13:27:29 crc kubenswrapper[4824]: W1124 13:27:29.643834 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda252dabf_79e7_442e_990c_3f15bc47d536.slice/crio-a0d06ac1751575fd84bde4396e64eb8171553b2302727a24b511eee4609d9b7c WatchSource:0}: Error finding container a0d06ac1751575fd84bde4396e64eb8171553b2302727a24b511eee4609d9b7c: Status 404 returned error can't find the container with id a0d06ac1751575fd84bde4396e64eb8171553b2302727a24b511eee4609d9b7c Nov 24 13:27:30 crc kubenswrapper[4824]: I1124 13:27:30.165132 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:30 crc kubenswrapper[4824]: I1124 13:27:30.211793 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:30 crc kubenswrapper[4824]: I1124 13:27:30.403690 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" event={"ID":"a252dabf-79e7-442e-990c-3f15bc47d536","Type":"ContainerStarted","Data":"a0d06ac1751575fd84bde4396e64eb8171553b2302727a24b511eee4609d9b7c"} Nov 24 13:27:30 crc kubenswrapper[4824]: I1124 13:27:30.406240 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6866949b48-flgn5" event={"ID":"eb457852-894e-4da1-aad8-c77c54da041b","Type":"ContainerStarted","Data":"3ae4bac116ab265a8a807621d054358c9edb5563b9adb12a8a381b1e8bccc0f4"} Nov 24 13:27:30 crc kubenswrapper[4824]: I1124 13:27:30.980225 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6866949b48-flgn5" podStartSLOduration=2.980208406 podStartE2EDuration="2.980208406s" podCreationTimestamp="2025-11-24 13:27:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:27:30.427572932 +0000 UTC m=+792.067112242" watchObservedRunningTime="2025-11-24 13:27:30.980208406 +0000 UTC m=+792.619747716" Nov 24 13:27:30 crc kubenswrapper[4824]: I1124 13:27:30.982196 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pjpp9"] Nov 24 13:27:31 crc kubenswrapper[4824]: I1124 13:27:31.413691 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gw4fj" event={"ID":"1e8318a0-5501-47b7-836a-4d650de94219","Type":"ContainerStarted","Data":"93f9895d27bf6ccf35f9ed1e1a516855528484bbb4fbd9d526229d16b0c6b68d"} Nov 24 13:27:31 crc kubenswrapper[4824]: I1124 13:27:31.414175 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:31 crc kubenswrapper[4824]: I1124 13:27:31.415856 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" event={"ID":"8876f622-9722-4257-8526-9f1f0e6a3daa","Type":"ContainerStarted","Data":"d257d0efa660f63e7e4fe1d6b114d9fcc2187f2090ea95534bbf645bde8dfea8"} Nov 24 13:27:31 crc kubenswrapper[4824]: I1124 13:27:31.415930 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:31 crc kubenswrapper[4824]: I1124 13:27:31.417306 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5" event={"ID":"2717ebec-b84b-4d07-a1ac-67209e7e5e32","Type":"ContainerStarted","Data":"9c50cf1981744ea4ec94fc662711798fb05258fd0d5f94982498ff086309400a"} Nov 24 13:27:31 crc kubenswrapper[4824]: I1124 13:27:31.417544 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pjpp9" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerName="registry-server" containerID="cri-o://00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863" gracePeriod=2 Nov 24 13:27:31 crc kubenswrapper[4824]: I1124 13:27:31.432909 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-gw4fj" podStartSLOduration=0.983880435 podStartE2EDuration="3.432892089s" podCreationTimestamp="2025-11-24 13:27:28 +0000 UTC" firstStartedPulling="2025-11-24 13:27:28.521779378 +0000 UTC m=+790.161318688" lastFinishedPulling="2025-11-24 13:27:30.970791032 +0000 UTC m=+792.610330342" observedRunningTime="2025-11-24 13:27:31.430222239 +0000 UTC m=+793.069761549" watchObservedRunningTime="2025-11-24 13:27:31.432892089 +0000 UTC m=+793.072431399" Nov 24 13:27:31 crc kubenswrapper[4824]: I1124 13:27:31.449236 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" podStartSLOduration=1.43708582 podStartE2EDuration="3.449217992s" podCreationTimestamp="2025-11-24 13:27:28 +0000 UTC" firstStartedPulling="2025-11-24 13:27:29.002837626 +0000 UTC m=+790.642376926" lastFinishedPulling="2025-11-24 13:27:31.014969778 +0000 UTC m=+792.654509098" observedRunningTime="2025-11-24 13:27:31.442434686 +0000 UTC m=+793.081974006" watchObservedRunningTime="2025-11-24 13:27:31.449217992 +0000 UTC m=+793.088757302" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.320215 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.412142 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-catalog-content\") pod \"c25af346-6bf0-4e3b-91f1-d213724477a0\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.412242 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7stf\" (UniqueName: \"kubernetes.io/projected/c25af346-6bf0-4e3b-91f1-d213724477a0-kube-api-access-z7stf\") pod \"c25af346-6bf0-4e3b-91f1-d213724477a0\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.412334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-utilities\") pod \"c25af346-6bf0-4e3b-91f1-d213724477a0\" (UID: \"c25af346-6bf0-4e3b-91f1-d213724477a0\") " Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.413369 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-utilities" (OuterVolumeSpecName: "utilities") pod "c25af346-6bf0-4e3b-91f1-d213724477a0" (UID: "c25af346-6bf0-4e3b-91f1-d213724477a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.415587 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c25af346-6bf0-4e3b-91f1-d213724477a0-kube-api-access-z7stf" (OuterVolumeSpecName: "kube-api-access-z7stf") pod "c25af346-6bf0-4e3b-91f1-d213724477a0" (UID: "c25af346-6bf0-4e3b-91f1-d213724477a0"). InnerVolumeSpecName "kube-api-access-z7stf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.422943 4824 generic.go:334] "Generic (PLEG): container finished" podID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerID="00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863" exitCode=0 Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.423013 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjpp9" event={"ID":"c25af346-6bf0-4e3b-91f1-d213724477a0","Type":"ContainerDied","Data":"00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863"} Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.423044 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjpp9" event={"ID":"c25af346-6bf0-4e3b-91f1-d213724477a0","Type":"ContainerDied","Data":"36ea8291abfe71038ef2895a8c5a690e1a2503aa3020cbd3af7c8e1aae7b639f"} Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.423061 4824 scope.go:117] "RemoveContainer" containerID="00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.423143 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pjpp9" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.464354 4824 scope.go:117] "RemoveContainer" containerID="87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.497045 4824 scope.go:117] "RemoveContainer" containerID="9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.513398 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7stf\" (UniqueName: \"kubernetes.io/projected/c25af346-6bf0-4e3b-91f1-d213724477a0-kube-api-access-z7stf\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.513430 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.522787 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c25af346-6bf0-4e3b-91f1-d213724477a0" (UID: "c25af346-6bf0-4e3b-91f1-d213724477a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.523270 4824 scope.go:117] "RemoveContainer" containerID="00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863" Nov 24 13:27:32 crc kubenswrapper[4824]: E1124 13:27:32.523827 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863\": container with ID starting with 00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863 not found: ID does not exist" containerID="00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.523919 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863"} err="failed to get container status \"00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863\": rpc error: code = NotFound desc = could not find container \"00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863\": container with ID starting with 00084666a4942cb792866c3f9ded5d547a02b264179ef7978305c83523187863 not found: ID does not exist" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.523948 4824 scope.go:117] "RemoveContainer" containerID="87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914" Nov 24 13:27:32 crc kubenswrapper[4824]: E1124 13:27:32.524488 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914\": container with ID starting with 87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914 not found: ID does not exist" containerID="87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.524522 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914"} err="failed to get container status \"87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914\": rpc error: code = NotFound desc = could not find container \"87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914\": container with ID starting with 87488c379fdfca4e04255c6eb54ff808f121599059cbd3843deb7808ba638914 not found: ID does not exist" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.524540 4824 scope.go:117] "RemoveContainer" containerID="9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9" Nov 24 13:27:32 crc kubenswrapper[4824]: E1124 13:27:32.524914 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9\": container with ID starting with 9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9 not found: ID does not exist" containerID="9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.524950 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9"} err="failed to get container status \"9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9\": rpc error: code = NotFound desc = could not find container \"9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9\": container with ID starting with 9f9674e9fc6b32df5f7d3f286ba83aa70422a4f812e998c85d262265009833b9 not found: ID does not exist" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.614084 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25af346-6bf0-4e3b-91f1-d213724477a0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.752689 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pjpp9"] Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.758825 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pjpp9"] Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.915753 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.915794 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:32 crc kubenswrapper[4824]: I1124 13:27:32.957144 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:33 crc kubenswrapper[4824]: I1124 13:27:33.017971 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" path="/var/lib/kubelet/pods/c25af346-6bf0-4e3b-91f1-d213724477a0/volumes" Nov 24 13:27:33 crc kubenswrapper[4824]: I1124 13:27:33.437183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" event={"ID":"a252dabf-79e7-442e-990c-3f15bc47d536","Type":"ContainerStarted","Data":"f5646b644f543b7014b178f656e854bb2dc7d24bdc39ecc23c645793039ffb41"} Nov 24 13:27:33 crc kubenswrapper[4824]: I1124 13:27:33.456458 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7sv5s" podStartSLOduration=2.75834092 podStartE2EDuration="5.456432456s" podCreationTimestamp="2025-11-24 13:27:28 +0000 UTC" firstStartedPulling="2025-11-24 13:27:29.645944827 +0000 UTC m=+791.285484137" lastFinishedPulling="2025-11-24 13:27:32.344036363 +0000 UTC m=+793.983575673" observedRunningTime="2025-11-24 13:27:33.448801378 +0000 UTC m=+795.088340688" watchObservedRunningTime="2025-11-24 13:27:33.456432456 +0000 UTC m=+795.095971766" Nov 24 13:27:33 crc kubenswrapper[4824]: I1124 13:27:33.519523 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:34 crc kubenswrapper[4824]: I1124 13:27:34.449895 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5" event={"ID":"2717ebec-b84b-4d07-a1ac-67209e7e5e32","Type":"ContainerStarted","Data":"87bf3da9835f27b0b52ba178086017032f48613129d13f82561b73541a055a94"} Nov 24 13:27:34 crc kubenswrapper[4824]: I1124 13:27:34.469993 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-gb5z5" podStartSLOduration=2.147338074 podStartE2EDuration="6.469977197s" podCreationTimestamp="2025-11-24 13:27:28 +0000 UTC" firstStartedPulling="2025-11-24 13:27:29.079856404 +0000 UTC m=+790.719395714" lastFinishedPulling="2025-11-24 13:27:33.402495517 +0000 UTC m=+795.042034837" observedRunningTime="2025-11-24 13:27:34.468114419 +0000 UTC m=+796.107653759" watchObservedRunningTime="2025-11-24 13:27:34.469977197 +0000 UTC m=+796.109516507" Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.381050 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jgrzs"] Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.453244 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jgrzs" podUID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerName="registry-server" containerID="cri-o://b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83" gracePeriod=2 Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.817375 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.860010 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2qhc\" (UniqueName: \"kubernetes.io/projected/9cab54fb-6f5d-4999-9223-d10de0017da3-kube-api-access-c2qhc\") pod \"9cab54fb-6f5d-4999-9223-d10de0017da3\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.860062 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-utilities\") pod \"9cab54fb-6f5d-4999-9223-d10de0017da3\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.860089 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-catalog-content\") pod \"9cab54fb-6f5d-4999-9223-d10de0017da3\" (UID: \"9cab54fb-6f5d-4999-9223-d10de0017da3\") " Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.861065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-utilities" (OuterVolumeSpecName: "utilities") pod "9cab54fb-6f5d-4999-9223-d10de0017da3" (UID: "9cab54fb-6f5d-4999-9223-d10de0017da3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.865139 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cab54fb-6f5d-4999-9223-d10de0017da3-kube-api-access-c2qhc" (OuterVolumeSpecName: "kube-api-access-c2qhc") pod "9cab54fb-6f5d-4999-9223-d10de0017da3" (UID: "9cab54fb-6f5d-4999-9223-d10de0017da3"). InnerVolumeSpecName "kube-api-access-c2qhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.878195 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cab54fb-6f5d-4999-9223-d10de0017da3" (UID: "9cab54fb-6f5d-4999-9223-d10de0017da3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.961547 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2qhc\" (UniqueName: \"kubernetes.io/projected/9cab54fb-6f5d-4999-9223-d10de0017da3-kube-api-access-c2qhc\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.961598 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:35 crc kubenswrapper[4824]: I1124 13:27:35.961616 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cab54fb-6f5d-4999-9223-d10de0017da3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.462698 4824 generic.go:334] "Generic (PLEG): container finished" podID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerID="b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83" exitCode=0 Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.462743 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jgrzs" event={"ID":"9cab54fb-6f5d-4999-9223-d10de0017da3","Type":"ContainerDied","Data":"b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83"} Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.462781 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jgrzs" event={"ID":"9cab54fb-6f5d-4999-9223-d10de0017da3","Type":"ContainerDied","Data":"de9186eaf4b0f245417290b024bbb9a75392552e76a4c26368037863a218aefc"} Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.462785 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jgrzs" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.462832 4824 scope.go:117] "RemoveContainer" containerID="b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.490271 4824 scope.go:117] "RemoveContainer" containerID="430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.512156 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jgrzs"] Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.516858 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jgrzs"] Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.518569 4824 scope.go:117] "RemoveContainer" containerID="924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.535878 4824 scope.go:117] "RemoveContainer" containerID="b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83" Nov 24 13:27:36 crc kubenswrapper[4824]: E1124 13:27:36.536195 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83\": container with ID starting with b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83 not found: ID does not exist" containerID="b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.536233 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83"} err="failed to get container status \"b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83\": rpc error: code = NotFound desc = could not find container \"b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83\": container with ID starting with b1962cc946e634cd023b3f765e5a988db8b5431a066f8564c318f69fe382ab83 not found: ID does not exist" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.536259 4824 scope.go:117] "RemoveContainer" containerID="430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f" Nov 24 13:27:36 crc kubenswrapper[4824]: E1124 13:27:36.536593 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f\": container with ID starting with 430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f not found: ID does not exist" containerID="430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.536612 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f"} err="failed to get container status \"430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f\": rpc error: code = NotFound desc = could not find container \"430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f\": container with ID starting with 430048f7630fd53e70d6c4fa5adcce712e8b756005176bbfdf0be2f0651ea12f not found: ID does not exist" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.536626 4824 scope.go:117] "RemoveContainer" containerID="924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9" Nov 24 13:27:36 crc kubenswrapper[4824]: E1124 13:27:36.536854 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9\": container with ID starting with 924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9 not found: ID does not exist" containerID="924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9" Nov 24 13:27:36 crc kubenswrapper[4824]: I1124 13:27:36.536874 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9"} err="failed to get container status \"924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9\": rpc error: code = NotFound desc = could not find container \"924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9\": container with ID starting with 924df593b365bb3a048f1512a00e3b0bff2618ab07486f56fc4b8f71309861e9 not found: ID does not exist" Nov 24 13:27:37 crc kubenswrapper[4824]: I1124 13:27:37.017650 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cab54fb-6f5d-4999-9223-d10de0017da3" path="/var/lib/kubelet/pods/9cab54fb-6f5d-4999-9223-d10de0017da3/volumes" Nov 24 13:27:38 crc kubenswrapper[4824]: I1124 13:27:38.441075 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-gw4fj" Nov 24 13:27:38 crc kubenswrapper[4824]: I1124 13:27:38.843402 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:38 crc kubenswrapper[4824]: I1124 13:27:38.843445 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:38 crc kubenswrapper[4824]: I1124 13:27:38.849248 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:39 crc kubenswrapper[4824]: I1124 13:27:39.482734 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6866949b48-flgn5" Nov 24 13:27:39 crc kubenswrapper[4824]: I1124 13:27:39.539261 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ll2hq"] Nov 24 13:27:40 crc kubenswrapper[4824]: I1124 13:27:40.787729 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:27:40 crc kubenswrapper[4824]: I1124 13:27:40.788946 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:27:48 crc kubenswrapper[4824]: I1124 13:27:48.369030 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-qs6l4" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.822557 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dxnrn"] Nov 24 13:27:57 crc kubenswrapper[4824]: E1124 13:27:57.823561 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerName="extract-content" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.823577 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerName="extract-content" Nov 24 13:27:57 crc kubenswrapper[4824]: E1124 13:27:57.823590 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerName="registry-server" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.823597 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerName="registry-server" Nov 24 13:27:57 crc kubenswrapper[4824]: E1124 13:27:57.823618 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerName="extract-utilities" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.823625 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerName="extract-utilities" Nov 24 13:27:57 crc kubenswrapper[4824]: E1124 13:27:57.823636 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerName="registry-server" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.823644 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerName="registry-server" Nov 24 13:27:57 crc kubenswrapper[4824]: E1124 13:27:57.823652 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerName="extract-utilities" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.823659 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerName="extract-utilities" Nov 24 13:27:57 crc kubenswrapper[4824]: E1124 13:27:57.823667 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerName="extract-content" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.823674 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerName="extract-content" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.823849 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cab54fb-6f5d-4999-9223-d10de0017da3" containerName="registry-server" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.823864 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c25af346-6bf0-4e3b-91f1-d213724477a0" containerName="registry-server" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.824695 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.841199 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dxnrn"] Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.946068 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk779\" (UniqueName: \"kubernetes.io/projected/509ed310-70b7-4fd2-9343-226c1d92b157-kube-api-access-sk779\") pod \"community-operators-dxnrn\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.946171 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-utilities\") pod \"community-operators-dxnrn\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:57 crc kubenswrapper[4824]: I1124 13:27:57.946229 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-catalog-content\") pod \"community-operators-dxnrn\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:58 crc kubenswrapper[4824]: I1124 13:27:58.047880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-catalog-content\") pod \"community-operators-dxnrn\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:58 crc kubenswrapper[4824]: I1124 13:27:58.047966 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk779\" (UniqueName: \"kubernetes.io/projected/509ed310-70b7-4fd2-9343-226c1d92b157-kube-api-access-sk779\") pod \"community-operators-dxnrn\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:58 crc kubenswrapper[4824]: I1124 13:27:58.048031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-utilities\") pod \"community-operators-dxnrn\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:58 crc kubenswrapper[4824]: I1124 13:27:58.048377 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-utilities\") pod \"community-operators-dxnrn\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:58 crc kubenswrapper[4824]: I1124 13:27:58.048379 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-catalog-content\") pod \"community-operators-dxnrn\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:58 crc kubenswrapper[4824]: I1124 13:27:58.069556 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk779\" (UniqueName: \"kubernetes.io/projected/509ed310-70b7-4fd2-9343-226c1d92b157-kube-api-access-sk779\") pod \"community-operators-dxnrn\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:58 crc kubenswrapper[4824]: I1124 13:27:58.153915 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:27:58 crc kubenswrapper[4824]: I1124 13:27:58.688485 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dxnrn"] Nov 24 13:27:59 crc kubenswrapper[4824]: I1124 13:27:59.602952 4824 generic.go:334] "Generic (PLEG): container finished" podID="509ed310-70b7-4fd2-9343-226c1d92b157" containerID="fe4de3ccaef10e1a36e1589cfb38db83643e194a12ee497d8290a9b0342722bc" exitCode=0 Nov 24 13:27:59 crc kubenswrapper[4824]: I1124 13:27:59.603003 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxnrn" event={"ID":"509ed310-70b7-4fd2-9343-226c1d92b157","Type":"ContainerDied","Data":"fe4de3ccaef10e1a36e1589cfb38db83643e194a12ee497d8290a9b0342722bc"} Nov 24 13:27:59 crc kubenswrapper[4824]: I1124 13:27:59.603300 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxnrn" event={"ID":"509ed310-70b7-4fd2-9343-226c1d92b157","Type":"ContainerStarted","Data":"0a35edceacb13cedc283e5570740091ad2541165ffca57f740608f04975ed9b4"} Nov 24 13:28:01 crc kubenswrapper[4824]: I1124 13:28:01.621344 4824 generic.go:334] "Generic (PLEG): container finished" podID="509ed310-70b7-4fd2-9343-226c1d92b157" containerID="4088529a063f1a29d744b42a02a77860990b17f084776682a3f024152eafa0cc" exitCode=0 Nov 24 13:28:01 crc kubenswrapper[4824]: I1124 13:28:01.621626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxnrn" event={"ID":"509ed310-70b7-4fd2-9343-226c1d92b157","Type":"ContainerDied","Data":"4088529a063f1a29d744b42a02a77860990b17f084776682a3f024152eafa0cc"} Nov 24 13:28:03 crc kubenswrapper[4824]: I1124 13:28:03.634868 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxnrn" event={"ID":"509ed310-70b7-4fd2-9343-226c1d92b157","Type":"ContainerStarted","Data":"2fd5efd933565ce5b81e81b0ec0c85f119f5ec48ab2e266facee14c2caeb265e"} Nov 24 13:28:03 crc kubenswrapper[4824]: I1124 13:28:03.657167 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dxnrn" podStartSLOduration=3.790580996 podStartE2EDuration="6.657150481s" podCreationTimestamp="2025-11-24 13:27:57 +0000 UTC" firstStartedPulling="2025-11-24 13:27:59.606970994 +0000 UTC m=+821.246510314" lastFinishedPulling="2025-11-24 13:28:02.473540489 +0000 UTC m=+824.113079799" observedRunningTime="2025-11-24 13:28:03.654591685 +0000 UTC m=+825.294130995" watchObservedRunningTime="2025-11-24 13:28:03.657150481 +0000 UTC m=+825.296689791" Nov 24 13:28:03 crc kubenswrapper[4824]: I1124 13:28:03.848144 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v"] Nov 24 13:28:03 crc kubenswrapper[4824]: I1124 13:28:03.849107 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:03 crc kubenswrapper[4824]: I1124 13:28:03.851654 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 13:28:03 crc kubenswrapper[4824]: I1124 13:28:03.864703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v"] Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.032057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vpxl\" (UniqueName: \"kubernetes.io/projected/0dbae9a3-1221-48ad-84a8-327d9f8176e7-kube-api-access-7vpxl\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.032120 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.032154 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.132880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vpxl\" (UniqueName: \"kubernetes.io/projected/0dbae9a3-1221-48ad-84a8-327d9f8176e7-kube-api-access-7vpxl\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.132924 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.132975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.133651 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.133716 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.161851 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vpxl\" (UniqueName: \"kubernetes.io/projected/0dbae9a3-1221-48ad-84a8-327d9f8176e7-kube-api-access-7vpxl\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.169385 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.576049 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-ll2hq" podUID="76c0d257-aca8-434e-97b2-067ddc1e7f66" containerName="console" containerID="cri-o://0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371" gracePeriod=15 Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.587183 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v"] Nov 24 13:28:04 crc kubenswrapper[4824]: W1124 13:28:04.594380 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dbae9a3_1221_48ad_84a8_327d9f8176e7.slice/crio-e6cde65bc359c8a52958b6b94cea36fa1dc1e751712711fceb5dda242c060113 WatchSource:0}: Error finding container e6cde65bc359c8a52958b6b94cea36fa1dc1e751712711fceb5dda242c060113: Status 404 returned error can't find the container with id e6cde65bc359c8a52958b6b94cea36fa1dc1e751712711fceb5dda242c060113 Nov 24 13:28:04 crc kubenswrapper[4824]: I1124 13:28:04.649062 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" event={"ID":"0dbae9a3-1221-48ad-84a8-327d9f8176e7","Type":"ContainerStarted","Data":"e6cde65bc359c8a52958b6b94cea36fa1dc1e751712711fceb5dda242c060113"} Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.193951 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ll2hq_76c0d257-aca8-434e-97b2-067ddc1e7f66/console/0.log" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.194088 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.354051 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-service-ca\") pod \"76c0d257-aca8-434e-97b2-067ddc1e7f66\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.354140 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-oauth-serving-cert\") pod \"76c0d257-aca8-434e-97b2-067ddc1e7f66\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.354197 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-oauth-config\") pod \"76c0d257-aca8-434e-97b2-067ddc1e7f66\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.354235 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-serving-cert\") pod \"76c0d257-aca8-434e-97b2-067ddc1e7f66\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.354289 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-config\") pod \"76c0d257-aca8-434e-97b2-067ddc1e7f66\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.354324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-trusted-ca-bundle\") pod \"76c0d257-aca8-434e-97b2-067ddc1e7f66\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.354379 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqsxl\" (UniqueName: \"kubernetes.io/projected/76c0d257-aca8-434e-97b2-067ddc1e7f66-kube-api-access-dqsxl\") pod \"76c0d257-aca8-434e-97b2-067ddc1e7f66\" (UID: \"76c0d257-aca8-434e-97b2-067ddc1e7f66\") " Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.355181 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-config" (OuterVolumeSpecName: "console-config") pod "76c0d257-aca8-434e-97b2-067ddc1e7f66" (UID: "76c0d257-aca8-434e-97b2-067ddc1e7f66"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.355247 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-service-ca" (OuterVolumeSpecName: "service-ca") pod "76c0d257-aca8-434e-97b2-067ddc1e7f66" (UID: "76c0d257-aca8-434e-97b2-067ddc1e7f66"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.355272 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "76c0d257-aca8-434e-97b2-067ddc1e7f66" (UID: "76c0d257-aca8-434e-97b2-067ddc1e7f66"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.355363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "76c0d257-aca8-434e-97b2-067ddc1e7f66" (UID: "76c0d257-aca8-434e-97b2-067ddc1e7f66"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.360062 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "76c0d257-aca8-434e-97b2-067ddc1e7f66" (UID: "76c0d257-aca8-434e-97b2-067ddc1e7f66"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.361023 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "76c0d257-aca8-434e-97b2-067ddc1e7f66" (UID: "76c0d257-aca8-434e-97b2-067ddc1e7f66"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.363205 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76c0d257-aca8-434e-97b2-067ddc1e7f66-kube-api-access-dqsxl" (OuterVolumeSpecName: "kube-api-access-dqsxl") pod "76c0d257-aca8-434e-97b2-067ddc1e7f66" (UID: "76c0d257-aca8-434e-97b2-067ddc1e7f66"). InnerVolumeSpecName "kube-api-access-dqsxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.455922 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqsxl\" (UniqueName: \"kubernetes.io/projected/76c0d257-aca8-434e-97b2-067ddc1e7f66-kube-api-access-dqsxl\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.455976 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.455985 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.455993 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.456001 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.456011 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.456019 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76c0d257-aca8-434e-97b2-067ddc1e7f66-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.659887 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ll2hq_76c0d257-aca8-434e-97b2-067ddc1e7f66/console/0.log" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.659973 4824 generic.go:334] "Generic (PLEG): container finished" podID="76c0d257-aca8-434e-97b2-067ddc1e7f66" containerID="0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371" exitCode=2 Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.660133 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ll2hq" event={"ID":"76c0d257-aca8-434e-97b2-067ddc1e7f66","Type":"ContainerDied","Data":"0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371"} Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.660202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ll2hq" event={"ID":"76c0d257-aca8-434e-97b2-067ddc1e7f66","Type":"ContainerDied","Data":"368c1773b03ecb2b9417586586be7b4aa39785ba79e860880bcada8e49fc0b04"} Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.660301 4824 scope.go:117] "RemoveContainer" containerID="0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.660324 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ll2hq" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.662773 4824 generic.go:334] "Generic (PLEG): container finished" podID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerID="0df0ec63c03d8ae6a2b9314f96f488df600d5e845ee6714cd539d3e1acb708d0" exitCode=0 Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.662947 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" event={"ID":"0dbae9a3-1221-48ad-84a8-327d9f8176e7","Type":"ContainerDied","Data":"0df0ec63c03d8ae6a2b9314f96f488df600d5e845ee6714cd539d3e1acb708d0"} Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.719952 4824 scope.go:117] "RemoveContainer" containerID="0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371" Nov 24 13:28:05 crc kubenswrapper[4824]: E1124 13:28:05.720794 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371\": container with ID starting with 0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371 not found: ID does not exist" containerID="0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.720874 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371"} err="failed to get container status \"0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371\": rpc error: code = NotFound desc = could not find container \"0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371\": container with ID starting with 0a01c6df283730ea8c7b946609a18cd1f31335583d12a888bde70056b9674371 not found: ID does not exist" Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.736149 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ll2hq"] Nov 24 13:28:05 crc kubenswrapper[4824]: I1124 13:28:05.744325 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-ll2hq"] Nov 24 13:28:07 crc kubenswrapper[4824]: I1124 13:28:07.019913 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76c0d257-aca8-434e-97b2-067ddc1e7f66" path="/var/lib/kubelet/pods/76c0d257-aca8-434e-97b2-067ddc1e7f66/volumes" Nov 24 13:28:07 crc kubenswrapper[4824]: I1124 13:28:07.680078 4824 generic.go:334] "Generic (PLEG): container finished" podID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerID="0b22d8390ee73d1a8d4254846960b8d3ea39eb500686f5178856888a31aa0362" exitCode=0 Nov 24 13:28:07 crc kubenswrapper[4824]: I1124 13:28:07.680122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" event={"ID":"0dbae9a3-1221-48ad-84a8-327d9f8176e7","Type":"ContainerDied","Data":"0b22d8390ee73d1a8d4254846960b8d3ea39eb500686f5178856888a31aa0362"} Nov 24 13:28:08 crc kubenswrapper[4824]: I1124 13:28:08.154911 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:28:08 crc kubenswrapper[4824]: I1124 13:28:08.155241 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:28:08 crc kubenswrapper[4824]: I1124 13:28:08.203402 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:28:08 crc kubenswrapper[4824]: I1124 13:28:08.691580 4824 generic.go:334] "Generic (PLEG): container finished" podID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerID="50ec09b18c0396fc0f9e9c0307f53660e504c240b993a2c78c4423c9992251b0" exitCode=0 Nov 24 13:28:08 crc kubenswrapper[4824]: I1124 13:28:08.692531 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" event={"ID":"0dbae9a3-1221-48ad-84a8-327d9f8176e7","Type":"ContainerDied","Data":"50ec09b18c0396fc0f9e9c0307f53660e504c240b993a2c78c4423c9992251b0"} Nov 24 13:28:08 crc kubenswrapper[4824]: I1124 13:28:08.766668 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:28:09 crc kubenswrapper[4824]: I1124 13:28:09.970175 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.136286 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vpxl\" (UniqueName: \"kubernetes.io/projected/0dbae9a3-1221-48ad-84a8-327d9f8176e7-kube-api-access-7vpxl\") pod \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.136342 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-bundle\") pod \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.136435 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-util\") pod \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\" (UID: \"0dbae9a3-1221-48ad-84a8-327d9f8176e7\") " Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.138936 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-bundle" (OuterVolumeSpecName: "bundle") pod "0dbae9a3-1221-48ad-84a8-327d9f8176e7" (UID: "0dbae9a3-1221-48ad-84a8-327d9f8176e7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.146027 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dbae9a3-1221-48ad-84a8-327d9f8176e7-kube-api-access-7vpxl" (OuterVolumeSpecName: "kube-api-access-7vpxl") pod "0dbae9a3-1221-48ad-84a8-327d9f8176e7" (UID: "0dbae9a3-1221-48ad-84a8-327d9f8176e7"). InnerVolumeSpecName "kube-api-access-7vpxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.169345 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-util" (OuterVolumeSpecName: "util") pod "0dbae9a3-1221-48ad-84a8-327d9f8176e7" (UID: "0dbae9a3-1221-48ad-84a8-327d9f8176e7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.239509 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vpxl\" (UniqueName: \"kubernetes.io/projected/0dbae9a3-1221-48ad-84a8-327d9f8176e7-kube-api-access-7vpxl\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.239556 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.239577 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dbae9a3-1221-48ad-84a8-327d9f8176e7-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.704012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" event={"ID":"0dbae9a3-1221-48ad-84a8-327d9f8176e7","Type":"ContainerDied","Data":"e6cde65bc359c8a52958b6b94cea36fa1dc1e751712711fceb5dda242c060113"} Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.704328 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6cde65bc359c8a52958b6b94cea36fa1dc1e751712711fceb5dda242c060113" Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.704066 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v" Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.787925 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:28:10 crc kubenswrapper[4824]: I1124 13:28:10.789566 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:28:11 crc kubenswrapper[4824]: I1124 13:28:11.404114 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dxnrn"] Nov 24 13:28:11 crc kubenswrapper[4824]: I1124 13:28:11.404590 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dxnrn" podUID="509ed310-70b7-4fd2-9343-226c1d92b157" containerName="registry-server" containerID="cri-o://2fd5efd933565ce5b81e81b0ec0c85f119f5ec48ab2e266facee14c2caeb265e" gracePeriod=2 Nov 24 13:28:11 crc kubenswrapper[4824]: I1124 13:28:11.711863 4824 generic.go:334] "Generic (PLEG): container finished" podID="509ed310-70b7-4fd2-9343-226c1d92b157" containerID="2fd5efd933565ce5b81e81b0ec0c85f119f5ec48ab2e266facee14c2caeb265e" exitCode=0 Nov 24 13:28:11 crc kubenswrapper[4824]: I1124 13:28:11.711921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxnrn" event={"ID":"509ed310-70b7-4fd2-9343-226c1d92b157","Type":"ContainerDied","Data":"2fd5efd933565ce5b81e81b0ec0c85f119f5ec48ab2e266facee14c2caeb265e"} Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.356968 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.473035 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-utilities\") pod \"509ed310-70b7-4fd2-9343-226c1d92b157\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.473132 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-catalog-content\") pod \"509ed310-70b7-4fd2-9343-226c1d92b157\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.473189 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk779\" (UniqueName: \"kubernetes.io/projected/509ed310-70b7-4fd2-9343-226c1d92b157-kube-api-access-sk779\") pod \"509ed310-70b7-4fd2-9343-226c1d92b157\" (UID: \"509ed310-70b7-4fd2-9343-226c1d92b157\") " Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.474120 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-utilities" (OuterVolumeSpecName: "utilities") pod "509ed310-70b7-4fd2-9343-226c1d92b157" (UID: "509ed310-70b7-4fd2-9343-226c1d92b157"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.477919 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/509ed310-70b7-4fd2-9343-226c1d92b157-kube-api-access-sk779" (OuterVolumeSpecName: "kube-api-access-sk779") pod "509ed310-70b7-4fd2-9343-226c1d92b157" (UID: "509ed310-70b7-4fd2-9343-226c1d92b157"). InnerVolumeSpecName "kube-api-access-sk779". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.575269 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.575658 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk779\" (UniqueName: \"kubernetes.io/projected/509ed310-70b7-4fd2-9343-226c1d92b157-kube-api-access-sk779\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.722471 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxnrn" event={"ID":"509ed310-70b7-4fd2-9343-226c1d92b157","Type":"ContainerDied","Data":"0a35edceacb13cedc283e5570740091ad2541165ffca57f740608f04975ed9b4"} Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.722936 4824 scope.go:117] "RemoveContainer" containerID="2fd5efd933565ce5b81e81b0ec0c85f119f5ec48ab2e266facee14c2caeb265e" Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.722535 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dxnrn" Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.745069 4824 scope.go:117] "RemoveContainer" containerID="4088529a063f1a29d744b42a02a77860990b17f084776682a3f024152eafa0cc" Nov 24 13:28:12 crc kubenswrapper[4824]: I1124 13:28:12.765582 4824 scope.go:117] "RemoveContainer" containerID="fe4de3ccaef10e1a36e1589cfb38db83643e194a12ee497d8290a9b0342722bc" Nov 24 13:28:13 crc kubenswrapper[4824]: I1124 13:28:13.667091 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "509ed310-70b7-4fd2-9343-226c1d92b157" (UID: "509ed310-70b7-4fd2-9343-226c1d92b157"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:28:13 crc kubenswrapper[4824]: I1124 13:28:13.689197 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/509ed310-70b7-4fd2-9343-226c1d92b157-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:13 crc kubenswrapper[4824]: I1124 13:28:13.966087 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dxnrn"] Nov 24 13:28:13 crc kubenswrapper[4824]: I1124 13:28:13.981558 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dxnrn"] Nov 24 13:28:15 crc kubenswrapper[4824]: I1124 13:28:15.017084 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="509ed310-70b7-4fd2-9343-226c1d92b157" path="/var/lib/kubelet/pods/509ed310-70b7-4fd2-9343-226c1d92b157/volumes" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.089966 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4"] Nov 24 13:28:22 crc kubenswrapper[4824]: E1124 13:28:22.090854 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="509ed310-70b7-4fd2-9343-226c1d92b157" containerName="registry-server" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.090873 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="509ed310-70b7-4fd2-9343-226c1d92b157" containerName="registry-server" Nov 24 13:28:22 crc kubenswrapper[4824]: E1124 13:28:22.090887 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerName="pull" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.090895 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerName="pull" Nov 24 13:28:22 crc kubenswrapper[4824]: E1124 13:28:22.090905 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerName="util" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.090912 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerName="util" Nov 24 13:28:22 crc kubenswrapper[4824]: E1124 13:28:22.090923 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="509ed310-70b7-4fd2-9343-226c1d92b157" containerName="extract-utilities" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.090932 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="509ed310-70b7-4fd2-9343-226c1d92b157" containerName="extract-utilities" Nov 24 13:28:22 crc kubenswrapper[4824]: E1124 13:28:22.090941 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerName="extract" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.090948 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerName="extract" Nov 24 13:28:22 crc kubenswrapper[4824]: E1124 13:28:22.090963 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c0d257-aca8-434e-97b2-067ddc1e7f66" containerName="console" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.090969 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c0d257-aca8-434e-97b2-067ddc1e7f66" containerName="console" Nov 24 13:28:22 crc kubenswrapper[4824]: E1124 13:28:22.090979 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="509ed310-70b7-4fd2-9343-226c1d92b157" containerName="extract-content" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.090987 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="509ed310-70b7-4fd2-9343-226c1d92b157" containerName="extract-content" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.091108 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dbae9a3-1221-48ad-84a8-327d9f8176e7" containerName="extract" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.091122 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="509ed310-70b7-4fd2-9343-226c1d92b157" containerName="registry-server" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.091134 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="76c0d257-aca8-434e-97b2-067ddc1e7f66" containerName="console" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.091604 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.094424 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.094654 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.094861 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.094982 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-v5tdb" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.095069 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.110032 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4"] Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.191345 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/99485513-6292-40ee-aede-2ef1d8566124-webhook-cert\") pod \"metallb-operator-controller-manager-b6cff747d-x5ng4\" (UID: \"99485513-6292-40ee-aede-2ef1d8566124\") " pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.191750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhz4p\" (UniqueName: \"kubernetes.io/projected/99485513-6292-40ee-aede-2ef1d8566124-kube-api-access-fhz4p\") pod \"metallb-operator-controller-manager-b6cff747d-x5ng4\" (UID: \"99485513-6292-40ee-aede-2ef1d8566124\") " pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.191788 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/99485513-6292-40ee-aede-2ef1d8566124-apiservice-cert\") pod \"metallb-operator-controller-manager-b6cff747d-x5ng4\" (UID: \"99485513-6292-40ee-aede-2ef1d8566124\") " pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.293210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhz4p\" (UniqueName: \"kubernetes.io/projected/99485513-6292-40ee-aede-2ef1d8566124-kube-api-access-fhz4p\") pod \"metallb-operator-controller-manager-b6cff747d-x5ng4\" (UID: \"99485513-6292-40ee-aede-2ef1d8566124\") " pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.293270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/99485513-6292-40ee-aede-2ef1d8566124-apiservice-cert\") pod \"metallb-operator-controller-manager-b6cff747d-x5ng4\" (UID: \"99485513-6292-40ee-aede-2ef1d8566124\") " pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.293300 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/99485513-6292-40ee-aede-2ef1d8566124-webhook-cert\") pod \"metallb-operator-controller-manager-b6cff747d-x5ng4\" (UID: \"99485513-6292-40ee-aede-2ef1d8566124\") " pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.298974 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/99485513-6292-40ee-aede-2ef1d8566124-apiservice-cert\") pod \"metallb-operator-controller-manager-b6cff747d-x5ng4\" (UID: \"99485513-6292-40ee-aede-2ef1d8566124\") " pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.310655 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/99485513-6292-40ee-aede-2ef1d8566124-webhook-cert\") pod \"metallb-operator-controller-manager-b6cff747d-x5ng4\" (UID: \"99485513-6292-40ee-aede-2ef1d8566124\") " pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.330458 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhz4p\" (UniqueName: \"kubernetes.io/projected/99485513-6292-40ee-aede-2ef1d8566124-kube-api-access-fhz4p\") pod \"metallb-operator-controller-manager-b6cff747d-x5ng4\" (UID: \"99485513-6292-40ee-aede-2ef1d8566124\") " pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.446223 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.465624 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567"] Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.466462 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.473499 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.473579 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-5kgvl" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.473607 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.494445 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567"] Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.597399 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07d6946c-c191-4a18-9903-194bdc00b4a0-apiservice-cert\") pod \"metallb-operator-webhook-server-ffbf5bdd6-fr567\" (UID: \"07d6946c-c191-4a18-9903-194bdc00b4a0\") " pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.597671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvgzb\" (UniqueName: \"kubernetes.io/projected/07d6946c-c191-4a18-9903-194bdc00b4a0-kube-api-access-dvgzb\") pod \"metallb-operator-webhook-server-ffbf5bdd6-fr567\" (UID: \"07d6946c-c191-4a18-9903-194bdc00b4a0\") " pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.597722 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07d6946c-c191-4a18-9903-194bdc00b4a0-webhook-cert\") pod \"metallb-operator-webhook-server-ffbf5bdd6-fr567\" (UID: \"07d6946c-c191-4a18-9903-194bdc00b4a0\") " pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.699818 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07d6946c-c191-4a18-9903-194bdc00b4a0-webhook-cert\") pod \"metallb-operator-webhook-server-ffbf5bdd6-fr567\" (UID: \"07d6946c-c191-4a18-9903-194bdc00b4a0\") " pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.699866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07d6946c-c191-4a18-9903-194bdc00b4a0-apiservice-cert\") pod \"metallb-operator-webhook-server-ffbf5bdd6-fr567\" (UID: \"07d6946c-c191-4a18-9903-194bdc00b4a0\") " pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.699928 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvgzb\" (UniqueName: \"kubernetes.io/projected/07d6946c-c191-4a18-9903-194bdc00b4a0-kube-api-access-dvgzb\") pod \"metallb-operator-webhook-server-ffbf5bdd6-fr567\" (UID: \"07d6946c-c191-4a18-9903-194bdc00b4a0\") " pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.706278 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07d6946c-c191-4a18-9903-194bdc00b4a0-apiservice-cert\") pod \"metallb-operator-webhook-server-ffbf5bdd6-fr567\" (UID: \"07d6946c-c191-4a18-9903-194bdc00b4a0\") " pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.706898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07d6946c-c191-4a18-9903-194bdc00b4a0-webhook-cert\") pod \"metallb-operator-webhook-server-ffbf5bdd6-fr567\" (UID: \"07d6946c-c191-4a18-9903-194bdc00b4a0\") " pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.722754 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvgzb\" (UniqueName: \"kubernetes.io/projected/07d6946c-c191-4a18-9903-194bdc00b4a0-kube-api-access-dvgzb\") pod \"metallb-operator-webhook-server-ffbf5bdd6-fr567\" (UID: \"07d6946c-c191-4a18-9903-194bdc00b4a0\") " pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.794042 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:22 crc kubenswrapper[4824]: I1124 13:28:22.794904 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4"] Nov 24 13:28:23 crc kubenswrapper[4824]: W1124 13:28:23.084922 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07d6946c_c191_4a18_9903_194bdc00b4a0.slice/crio-2d206dbcd960ff6a37805a58d7bcc1ca97a2489da4c3c5b2acc73a06d2881101 WatchSource:0}: Error finding container 2d206dbcd960ff6a37805a58d7bcc1ca97a2489da4c3c5b2acc73a06d2881101: Status 404 returned error can't find the container with id 2d206dbcd960ff6a37805a58d7bcc1ca97a2489da4c3c5b2acc73a06d2881101 Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.086921 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567"] Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.607784 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8kv5l"] Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.609130 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.625947 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8kv5l"] Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.713400 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdrzx\" (UniqueName: \"kubernetes.io/projected/a0f1ca4f-6cd2-4898-bfea-e20593381658-kube-api-access-cdrzx\") pod \"certified-operators-8kv5l\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.713471 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-catalog-content\") pod \"certified-operators-8kv5l\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.713516 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-utilities\") pod \"certified-operators-8kv5l\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.786174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" event={"ID":"07d6946c-c191-4a18-9903-194bdc00b4a0","Type":"ContainerStarted","Data":"2d206dbcd960ff6a37805a58d7bcc1ca97a2489da4c3c5b2acc73a06d2881101"} Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.787297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" event={"ID":"99485513-6292-40ee-aede-2ef1d8566124","Type":"ContainerStarted","Data":"fed902d667cf47080057e2b7790c50fd683aca8906c0746f7bc5413a29b59910"} Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.815780 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdrzx\" (UniqueName: \"kubernetes.io/projected/a0f1ca4f-6cd2-4898-bfea-e20593381658-kube-api-access-cdrzx\") pod \"certified-operators-8kv5l\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.815849 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-catalog-content\") pod \"certified-operators-8kv5l\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.815888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-utilities\") pod \"certified-operators-8kv5l\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.816369 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-catalog-content\") pod \"certified-operators-8kv5l\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.816388 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-utilities\") pod \"certified-operators-8kv5l\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.836614 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdrzx\" (UniqueName: \"kubernetes.io/projected/a0f1ca4f-6cd2-4898-bfea-e20593381658-kube-api-access-cdrzx\") pod \"certified-operators-8kv5l\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:23 crc kubenswrapper[4824]: I1124 13:28:23.938180 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:24 crc kubenswrapper[4824]: I1124 13:28:24.205328 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8kv5l"] Nov 24 13:28:24 crc kubenswrapper[4824]: I1124 13:28:24.794981 4824 generic.go:334] "Generic (PLEG): container finished" podID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerID="76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba" exitCode=0 Nov 24 13:28:24 crc kubenswrapper[4824]: I1124 13:28:24.795020 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kv5l" event={"ID":"a0f1ca4f-6cd2-4898-bfea-e20593381658","Type":"ContainerDied","Data":"76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba"} Nov 24 13:28:24 crc kubenswrapper[4824]: I1124 13:28:24.795047 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kv5l" event={"ID":"a0f1ca4f-6cd2-4898-bfea-e20593381658","Type":"ContainerStarted","Data":"cb7dd0e12215d25b79e3b86a9d3111c188a15b5957de62f35c4e48bebae721cb"} Nov 24 13:28:31 crc kubenswrapper[4824]: I1124 13:28:31.834239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" event={"ID":"99485513-6292-40ee-aede-2ef1d8566124","Type":"ContainerStarted","Data":"674689f54681b08b3f9f0a487592e9c76591a2e2d578c5ea73c60fc823b69fad"} Nov 24 13:28:31 crc kubenswrapper[4824]: I1124 13:28:31.834961 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:28:31 crc kubenswrapper[4824]: I1124 13:28:31.837560 4824 generic.go:334] "Generic (PLEG): container finished" podID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerID="f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003" exitCode=0 Nov 24 13:28:31 crc kubenswrapper[4824]: I1124 13:28:31.837783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kv5l" event={"ID":"a0f1ca4f-6cd2-4898-bfea-e20593381658","Type":"ContainerDied","Data":"f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003"} Nov 24 13:28:31 crc kubenswrapper[4824]: I1124 13:28:31.839926 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" event={"ID":"07d6946c-c191-4a18-9903-194bdc00b4a0","Type":"ContainerStarted","Data":"4a88a01d9e9f5fa61751a3071e8c094ade508a97f0a420e247f5c5812d631cc7"} Nov 24 13:28:31 crc kubenswrapper[4824]: I1124 13:28:31.840285 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:31 crc kubenswrapper[4824]: I1124 13:28:31.858701 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" podStartSLOduration=1.643554649 podStartE2EDuration="9.858680029s" podCreationTimestamp="2025-11-24 13:28:22 +0000 UTC" firstStartedPulling="2025-11-24 13:28:22.811386352 +0000 UTC m=+844.450925662" lastFinishedPulling="2025-11-24 13:28:31.026511732 +0000 UTC m=+852.666051042" observedRunningTime="2025-11-24 13:28:31.853301508 +0000 UTC m=+853.492840818" watchObservedRunningTime="2025-11-24 13:28:31.858680029 +0000 UTC m=+853.498219339" Nov 24 13:28:31 crc kubenswrapper[4824]: I1124 13:28:31.881629 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" podStartSLOduration=1.913433853 podStartE2EDuration="9.881606669s" podCreationTimestamp="2025-11-24 13:28:22 +0000 UTC" firstStartedPulling="2025-11-24 13:28:23.0882701 +0000 UTC m=+844.727809410" lastFinishedPulling="2025-11-24 13:28:31.056442916 +0000 UTC m=+852.695982226" observedRunningTime="2025-11-24 13:28:31.880673265 +0000 UTC m=+853.520212585" watchObservedRunningTime="2025-11-24 13:28:31.881606669 +0000 UTC m=+853.521145989" Nov 24 13:28:32 crc kubenswrapper[4824]: I1124 13:28:32.847960 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kv5l" event={"ID":"a0f1ca4f-6cd2-4898-bfea-e20593381658","Type":"ContainerStarted","Data":"6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b"} Nov 24 13:28:32 crc kubenswrapper[4824]: I1124 13:28:32.865420 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8kv5l" podStartSLOduration=2.192318734 podStartE2EDuration="9.865399105s" podCreationTimestamp="2025-11-24 13:28:23 +0000 UTC" firstStartedPulling="2025-11-24 13:28:24.799166581 +0000 UTC m=+846.438705891" lastFinishedPulling="2025-11-24 13:28:32.472246952 +0000 UTC m=+854.111786262" observedRunningTime="2025-11-24 13:28:32.864370538 +0000 UTC m=+854.503909848" watchObservedRunningTime="2025-11-24 13:28:32.865399105 +0000 UTC m=+854.504938415" Nov 24 13:28:33 crc kubenswrapper[4824]: I1124 13:28:33.939085 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:33 crc kubenswrapper[4824]: I1124 13:28:33.939464 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:35 crc kubenswrapper[4824]: I1124 13:28:35.002900 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8kv5l" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerName="registry-server" probeResult="failure" output=< Nov 24 13:28:35 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 13:28:35 crc kubenswrapper[4824]: > Nov 24 13:28:40 crc kubenswrapper[4824]: I1124 13:28:40.788061 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:28:40 crc kubenswrapper[4824]: I1124 13:28:40.788592 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:28:40 crc kubenswrapper[4824]: I1124 13:28:40.788645 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:28:40 crc kubenswrapper[4824]: I1124 13:28:40.789314 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e56c6ac43b570680166358738b704b70ae28054c60648850ac0d5d06e92a00bf"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:28:40 crc kubenswrapper[4824]: I1124 13:28:40.789380 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://e56c6ac43b570680166358738b704b70ae28054c60648850ac0d5d06e92a00bf" gracePeriod=600 Nov 24 13:28:41 crc kubenswrapper[4824]: I1124 13:28:41.897463 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="e56c6ac43b570680166358738b704b70ae28054c60648850ac0d5d06e92a00bf" exitCode=0 Nov 24 13:28:41 crc kubenswrapper[4824]: I1124 13:28:41.897521 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"e56c6ac43b570680166358738b704b70ae28054c60648850ac0d5d06e92a00bf"} Nov 24 13:28:41 crc kubenswrapper[4824]: I1124 13:28:41.897994 4824 scope.go:117] "RemoveContainer" containerID="e631cd0c8c7e50628259c7f81b9d3169967909fa981a7ac6f1bca6bbd6fcf373" Nov 24 13:28:41 crc kubenswrapper[4824]: I1124 13:28:41.898007 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"62c6a3d914058b43356f45839d76c3c6a0c31d95c968d9b2fa78dd4dbda74ea8"} Nov 24 13:28:42 crc kubenswrapper[4824]: I1124 13:28:42.798032 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-ffbf5bdd6-fr567" Nov 24 13:28:43 crc kubenswrapper[4824]: I1124 13:28:43.988310 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:44 crc kubenswrapper[4824]: I1124 13:28:44.040258 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.316497 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8kv5l"] Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.316761 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8kv5l" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerName="registry-server" containerID="cri-o://6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b" gracePeriod=2 Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.781520 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.913980 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdrzx\" (UniqueName: \"kubernetes.io/projected/a0f1ca4f-6cd2-4898-bfea-e20593381658-kube-api-access-cdrzx\") pod \"a0f1ca4f-6cd2-4898-bfea-e20593381658\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.914064 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-catalog-content\") pod \"a0f1ca4f-6cd2-4898-bfea-e20593381658\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.914102 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-utilities\") pod \"a0f1ca4f-6cd2-4898-bfea-e20593381658\" (UID: \"a0f1ca4f-6cd2-4898-bfea-e20593381658\") " Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.915073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-utilities" (OuterVolumeSpecName: "utilities") pod "a0f1ca4f-6cd2-4898-bfea-e20593381658" (UID: "a0f1ca4f-6cd2-4898-bfea-e20593381658"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.915365 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.925790 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f1ca4f-6cd2-4898-bfea-e20593381658-kube-api-access-cdrzx" (OuterVolumeSpecName: "kube-api-access-cdrzx") pod "a0f1ca4f-6cd2-4898-bfea-e20593381658" (UID: "a0f1ca4f-6cd2-4898-bfea-e20593381658"). InnerVolumeSpecName "kube-api-access-cdrzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.929364 4824 generic.go:334] "Generic (PLEG): container finished" podID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerID="6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b" exitCode=0 Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.929395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kv5l" event={"ID":"a0f1ca4f-6cd2-4898-bfea-e20593381658","Type":"ContainerDied","Data":"6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b"} Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.929421 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kv5l" event={"ID":"a0f1ca4f-6cd2-4898-bfea-e20593381658","Type":"ContainerDied","Data":"cb7dd0e12215d25b79e3b86a9d3111c188a15b5957de62f35c4e48bebae721cb"} Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.929448 4824 scope.go:117] "RemoveContainer" containerID="6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.929551 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kv5l" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.952262 4824 scope.go:117] "RemoveContainer" containerID="f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.969356 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0f1ca4f-6cd2-4898-bfea-e20593381658" (UID: "a0f1ca4f-6cd2-4898-bfea-e20593381658"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.978637 4824 scope.go:117] "RemoveContainer" containerID="76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.990798 4824 scope.go:117] "RemoveContainer" containerID="6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b" Nov 24 13:28:46 crc kubenswrapper[4824]: E1124 13:28:46.991296 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b\": container with ID starting with 6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b not found: ID does not exist" containerID="6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.991335 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b"} err="failed to get container status \"6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b\": rpc error: code = NotFound desc = could not find container \"6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b\": container with ID starting with 6600d40f4e12abe82653ef3867e4a40fa1bc2af4357dcca38f9708190372b42b not found: ID does not exist" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.991361 4824 scope.go:117] "RemoveContainer" containerID="f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003" Nov 24 13:28:46 crc kubenswrapper[4824]: E1124 13:28:46.992251 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003\": container with ID starting with f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003 not found: ID does not exist" containerID="f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.992308 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003"} err="failed to get container status \"f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003\": rpc error: code = NotFound desc = could not find container \"f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003\": container with ID starting with f09d0b26295f71daaedd1b4ea29fe58b214a98932e76c910cef346f292f94003 not found: ID does not exist" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.992336 4824 scope.go:117] "RemoveContainer" containerID="76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba" Nov 24 13:28:46 crc kubenswrapper[4824]: E1124 13:28:46.992664 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba\": container with ID starting with 76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba not found: ID does not exist" containerID="76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba" Nov 24 13:28:46 crc kubenswrapper[4824]: I1124 13:28:46.992695 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba"} err="failed to get container status \"76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba\": rpc error: code = NotFound desc = could not find container \"76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba\": container with ID starting with 76c1486052e28db35c0b79341a74f7f65bb0871525c53935b166ad9796d55bba not found: ID does not exist" Nov 24 13:28:47 crc kubenswrapper[4824]: I1124 13:28:47.016059 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdrzx\" (UniqueName: \"kubernetes.io/projected/a0f1ca4f-6cd2-4898-bfea-e20593381658-kube-api-access-cdrzx\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:47 crc kubenswrapper[4824]: I1124 13:28:47.016083 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f1ca4f-6cd2-4898-bfea-e20593381658-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:28:47 crc kubenswrapper[4824]: I1124 13:28:47.243709 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8kv5l"] Nov 24 13:28:47 crc kubenswrapper[4824]: I1124 13:28:47.249735 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8kv5l"] Nov 24 13:28:49 crc kubenswrapper[4824]: I1124 13:28:49.019185 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" path="/var/lib/kubelet/pods/a0f1ca4f-6cd2-4898-bfea-e20593381658/volumes" Nov 24 13:29:02 crc kubenswrapper[4824]: I1124 13:29:02.448661 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-b6cff747d-x5ng4" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.169646 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-dxtg8"] Nov 24 13:29:03 crc kubenswrapper[4824]: E1124 13:29:03.169942 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerName="extract-utilities" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.169958 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerName="extract-utilities" Nov 24 13:29:03 crc kubenswrapper[4824]: E1124 13:29:03.169969 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerName="registry-server" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.169976 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerName="registry-server" Nov 24 13:29:03 crc kubenswrapper[4824]: E1124 13:29:03.169994 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerName="extract-content" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.170001 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerName="extract-content" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.170146 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f1ca4f-6cd2-4898-bfea-e20593381658" containerName="registry-server" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.172500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.175281 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q"] Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.176171 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.182573 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.182603 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.182574 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.182878 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-767h7" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.193243 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q"] Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.305784 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-r5w24"] Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.306603 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.312507 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.312526 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.314077 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.316587 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-frr-conf\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.316648 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d7ab142b-a25e-4591-9672-6209bf35a3a8-frr-startup\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.316691 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-metrics\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.316731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-frr-sockets\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.316758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8lp2\" (UniqueName: \"kubernetes.io/projected/db503b31-f22b-4b5c-a20f-970c15435851-kube-api-access-c8lp2\") pod \"frr-k8s-webhook-server-6998585d5-2gp4q\" (UID: \"db503b31-f22b-4b5c-a20f-970c15435851\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.316785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hvhk\" (UniqueName: \"kubernetes.io/projected/d7ab142b-a25e-4591-9672-6209bf35a3a8-kube-api-access-8hvhk\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.316834 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d7ab142b-a25e-4591-9672-6209bf35a3a8-metrics-certs\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.316869 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db503b31-f22b-4b5c-a20f-970c15435851-cert\") pod \"frr-k8s-webhook-server-6998585d5-2gp4q\" (UID: \"db503b31-f22b-4b5c-a20f-970c15435851\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.317305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-reloader\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.318910 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-npmzc" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.331638 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-m4d79"] Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.332444 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.334191 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.352684 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-m4d79"] Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-reloader\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-metrics-certs\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5fzm\" (UniqueName: \"kubernetes.io/projected/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-kube-api-access-s5fzm\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418859 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-frr-conf\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418885 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d7ab142b-a25e-4591-9672-6209bf35a3a8-frr-startup\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-memberlist\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-metrics\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-frr-sockets\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418976 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8lp2\" (UniqueName: \"kubernetes.io/projected/db503b31-f22b-4b5c-a20f-970c15435851-kube-api-access-c8lp2\") pod \"frr-k8s-webhook-server-6998585d5-2gp4q\" (UID: \"db503b31-f22b-4b5c-a20f-970c15435851\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.418998 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hvhk\" (UniqueName: \"kubernetes.io/projected/d7ab142b-a25e-4591-9672-6209bf35a3a8-kube-api-access-8hvhk\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.419031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d7ab142b-a25e-4591-9672-6209bf35a3a8-metrics-certs\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.419063 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db503b31-f22b-4b5c-a20f-970c15435851-cert\") pod \"frr-k8s-webhook-server-6998585d5-2gp4q\" (UID: \"db503b31-f22b-4b5c-a20f-970c15435851\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.419085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-metallb-excludel2\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.420214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-reloader\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.420306 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-frr-sockets\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.423196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-metrics\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.423534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d7ab142b-a25e-4591-9672-6209bf35a3a8-frr-startup\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.423675 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d7ab142b-a25e-4591-9672-6209bf35a3a8-frr-conf\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.432539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db503b31-f22b-4b5c-a20f-970c15435851-cert\") pod \"frr-k8s-webhook-server-6998585d5-2gp4q\" (UID: \"db503b31-f22b-4b5c-a20f-970c15435851\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.438622 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8lp2\" (UniqueName: \"kubernetes.io/projected/db503b31-f22b-4b5c-a20f-970c15435851-kube-api-access-c8lp2\") pod \"frr-k8s-webhook-server-6998585d5-2gp4q\" (UID: \"db503b31-f22b-4b5c-a20f-970c15435851\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.440833 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hvhk\" (UniqueName: \"kubernetes.io/projected/d7ab142b-a25e-4591-9672-6209bf35a3a8-kube-api-access-8hvhk\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.446343 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d7ab142b-a25e-4591-9672-6209bf35a3a8-metrics-certs\") pod \"frr-k8s-dxtg8\" (UID: \"d7ab142b-a25e-4591-9672-6209bf35a3a8\") " pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.502855 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.511866 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.520452 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-memberlist\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.520499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/077fe5ab-ea92-4aa8-8b76-dc791ae6d41b-cert\") pod \"controller-6c7b4b5f48-m4d79\" (UID: \"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b\") " pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.520548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/077fe5ab-ea92-4aa8-8b76-dc791ae6d41b-metrics-certs\") pod \"controller-6c7b4b5f48-m4d79\" (UID: \"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b\") " pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.520581 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-metallb-excludel2\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: E1124 13:29:03.520630 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 13:29:03 crc kubenswrapper[4824]: E1124 13:29:03.520696 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-memberlist podName:e2bbe7bd-470b-4363-a85c-7a2084da3e4e nodeName:}" failed. No retries permitted until 2025-11-24 13:29:04.020680054 +0000 UTC m=+885.660219364 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-memberlist") pod "speaker-r5w24" (UID: "e2bbe7bd-470b-4363-a85c-7a2084da3e4e") : secret "metallb-memberlist" not found Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.520864 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rsrl\" (UniqueName: \"kubernetes.io/projected/077fe5ab-ea92-4aa8-8b76-dc791ae6d41b-kube-api-access-5rsrl\") pod \"controller-6c7b4b5f48-m4d79\" (UID: \"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b\") " pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.520953 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-metrics-certs\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.521000 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5fzm\" (UniqueName: \"kubernetes.io/projected/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-kube-api-access-s5fzm\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.521336 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-metallb-excludel2\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.524373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-metrics-certs\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.544449 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5fzm\" (UniqueName: \"kubernetes.io/projected/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-kube-api-access-s5fzm\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.621958 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rsrl\" (UniqueName: \"kubernetes.io/projected/077fe5ab-ea92-4aa8-8b76-dc791ae6d41b-kube-api-access-5rsrl\") pod \"controller-6c7b4b5f48-m4d79\" (UID: \"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b\") " pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.622083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/077fe5ab-ea92-4aa8-8b76-dc791ae6d41b-cert\") pod \"controller-6c7b4b5f48-m4d79\" (UID: \"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b\") " pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.622148 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/077fe5ab-ea92-4aa8-8b76-dc791ae6d41b-metrics-certs\") pod \"controller-6c7b4b5f48-m4d79\" (UID: \"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b\") " pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.626990 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/077fe5ab-ea92-4aa8-8b76-dc791ae6d41b-cert\") pod \"controller-6c7b4b5f48-m4d79\" (UID: \"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b\") " pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.627401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/077fe5ab-ea92-4aa8-8b76-dc791ae6d41b-metrics-certs\") pod \"controller-6c7b4b5f48-m4d79\" (UID: \"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b\") " pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.643431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rsrl\" (UniqueName: \"kubernetes.io/projected/077fe5ab-ea92-4aa8-8b76-dc791ae6d41b-kube-api-access-5rsrl\") pod \"controller-6c7b4b5f48-m4d79\" (UID: \"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b\") " pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.655882 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:03 crc kubenswrapper[4824]: I1124 13:29:03.973393 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q"] Nov 24 13:29:04 crc kubenswrapper[4824]: I1124 13:29:04.027767 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-memberlist\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:04 crc kubenswrapper[4824]: E1124 13:29:04.027960 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 13:29:04 crc kubenswrapper[4824]: E1124 13:29:04.028009 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-memberlist podName:e2bbe7bd-470b-4363-a85c-7a2084da3e4e nodeName:}" failed. No retries permitted until 2025-11-24 13:29:05.027994716 +0000 UTC m=+886.667534026 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-memberlist") pod "speaker-r5w24" (UID: "e2bbe7bd-470b-4363-a85c-7a2084da3e4e") : secret "metallb-memberlist" not found Nov 24 13:29:04 crc kubenswrapper[4824]: I1124 13:29:04.053925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" event={"ID":"db503b31-f22b-4b5c-a20f-970c15435851","Type":"ContainerStarted","Data":"ed32350bdb9252c4fb1a9e7d4136cc93b1c9aa6dfefdfb9e0e94affaa937bb16"} Nov 24 13:29:04 crc kubenswrapper[4824]: I1124 13:29:04.054994 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerStarted","Data":"0fbf63f3dd70c0d33be82bfca3042e825de3012ed0b58fa1e8c34567982746bb"} Nov 24 13:29:04 crc kubenswrapper[4824]: I1124 13:29:04.080062 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-m4d79"] Nov 24 13:29:04 crc kubenswrapper[4824]: W1124 13:29:04.090413 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod077fe5ab_ea92_4aa8_8b76_dc791ae6d41b.slice/crio-5f6b91391b281d0b89ab64c343fe1a7636656ee5f616c9064e98c0886b3e36cf WatchSource:0}: Error finding container 5f6b91391b281d0b89ab64c343fe1a7636656ee5f616c9064e98c0886b3e36cf: Status 404 returned error can't find the container with id 5f6b91391b281d0b89ab64c343fe1a7636656ee5f616c9064e98c0886b3e36cf Nov 24 13:29:05 crc kubenswrapper[4824]: I1124 13:29:05.040252 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-memberlist\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:05 crc kubenswrapper[4824]: I1124 13:29:05.047088 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e2bbe7bd-470b-4363-a85c-7a2084da3e4e-memberlist\") pod \"speaker-r5w24\" (UID: \"e2bbe7bd-470b-4363-a85c-7a2084da3e4e\") " pod="metallb-system/speaker-r5w24" Nov 24 13:29:05 crc kubenswrapper[4824]: I1124 13:29:05.084558 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-m4d79" event={"ID":"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b","Type":"ContainerStarted","Data":"8438e48f721ea7c849e18af14ec62905d7b99663d2c737ae00a38276c2e85e8a"} Nov 24 13:29:05 crc kubenswrapper[4824]: I1124 13:29:05.084602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-m4d79" event={"ID":"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b","Type":"ContainerStarted","Data":"3908c0a6350ccbc2a45a9ce3e3aee8d5c37d80c9273e71912f69b21b0cb8f2e7"} Nov 24 13:29:05 crc kubenswrapper[4824]: I1124 13:29:05.084616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-m4d79" event={"ID":"077fe5ab-ea92-4aa8-8b76-dc791ae6d41b","Type":"ContainerStarted","Data":"5f6b91391b281d0b89ab64c343fe1a7636656ee5f616c9064e98c0886b3e36cf"} Nov 24 13:29:05 crc kubenswrapper[4824]: I1124 13:29:05.085560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:05 crc kubenswrapper[4824]: I1124 13:29:05.120641 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r5w24" Nov 24 13:29:05 crc kubenswrapper[4824]: I1124 13:29:05.121312 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-m4d79" podStartSLOduration=2.121299869 podStartE2EDuration="2.121299869s" podCreationTimestamp="2025-11-24 13:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:29:05.116600116 +0000 UTC m=+886.756139426" watchObservedRunningTime="2025-11-24 13:29:05.121299869 +0000 UTC m=+886.760839179" Nov 24 13:29:05 crc kubenswrapper[4824]: W1124 13:29:05.201427 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2bbe7bd_470b_4363_a85c_7a2084da3e4e.slice/crio-e57a7dc66a633d3800daa5eced12c7d36ae9fd4bf6cb14f4e3db1f4772114097 WatchSource:0}: Error finding container e57a7dc66a633d3800daa5eced12c7d36ae9fd4bf6cb14f4e3db1f4772114097: Status 404 returned error can't find the container with id e57a7dc66a633d3800daa5eced12c7d36ae9fd4bf6cb14f4e3db1f4772114097 Nov 24 13:29:06 crc kubenswrapper[4824]: I1124 13:29:06.103113 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r5w24" event={"ID":"e2bbe7bd-470b-4363-a85c-7a2084da3e4e","Type":"ContainerStarted","Data":"1e8b5603c50bc31bf20b68666e1cc857f401f3992a17523dfcd2f840813ada81"} Nov 24 13:29:06 crc kubenswrapper[4824]: I1124 13:29:06.103174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r5w24" event={"ID":"e2bbe7bd-470b-4363-a85c-7a2084da3e4e","Type":"ContainerStarted","Data":"71e591e9fa932b408f2cc834c4d65b3ff532567ab39b8e57aed5784c9c7aae6e"} Nov 24 13:29:06 crc kubenswrapper[4824]: I1124 13:29:06.103188 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r5w24" event={"ID":"e2bbe7bd-470b-4363-a85c-7a2084da3e4e","Type":"ContainerStarted","Data":"e57a7dc66a633d3800daa5eced12c7d36ae9fd4bf6cb14f4e3db1f4772114097"} Nov 24 13:29:06 crc kubenswrapper[4824]: I1124 13:29:06.103367 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-r5w24" Nov 24 13:29:09 crc kubenswrapper[4824]: I1124 13:29:09.047103 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-r5w24" podStartSLOduration=6.047085946 podStartE2EDuration="6.047085946s" podCreationTimestamp="2025-11-24 13:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:29:06.137512302 +0000 UTC m=+887.777051612" watchObservedRunningTime="2025-11-24 13:29:09.047085946 +0000 UTC m=+890.686625256" Nov 24 13:29:12 crc kubenswrapper[4824]: I1124 13:29:12.139708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" event={"ID":"db503b31-f22b-4b5c-a20f-970c15435851","Type":"ContainerStarted","Data":"45ebe8d09ba09551bbc079fe07e17a5d2bc9b295625e1c44c4c4d5b2fd0b313e"} Nov 24 13:29:12 crc kubenswrapper[4824]: I1124 13:29:12.140404 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:12 crc kubenswrapper[4824]: I1124 13:29:12.141173 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7ab142b-a25e-4591-9672-6209bf35a3a8" containerID="9666ce01ee7715c749ef31716499a9e470aa9993fb32926759e7bea641e42f22" exitCode=0 Nov 24 13:29:12 crc kubenswrapper[4824]: I1124 13:29:12.141202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerDied","Data":"9666ce01ee7715c749ef31716499a9e470aa9993fb32926759e7bea641e42f22"} Nov 24 13:29:12 crc kubenswrapper[4824]: I1124 13:29:12.203608 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" podStartSLOduration=1.7591992589999998 podStartE2EDuration="9.203588903s" podCreationTimestamp="2025-11-24 13:29:03 +0000 UTC" firstStartedPulling="2025-11-24 13:29:03.982454623 +0000 UTC m=+885.621993933" lastFinishedPulling="2025-11-24 13:29:11.426844267 +0000 UTC m=+893.066383577" observedRunningTime="2025-11-24 13:29:12.168444863 +0000 UTC m=+893.807984173" watchObservedRunningTime="2025-11-24 13:29:12.203588903 +0000 UTC m=+893.843128213" Nov 24 13:29:13 crc kubenswrapper[4824]: I1124 13:29:13.151184 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7ab142b-a25e-4591-9672-6209bf35a3a8" containerID="f183f0fefc1e2618714e70b7bcbd1512f8822b5db981352be236f79bb1db54b3" exitCode=0 Nov 24 13:29:13 crc kubenswrapper[4824]: I1124 13:29:13.151279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerDied","Data":"f183f0fefc1e2618714e70b7bcbd1512f8822b5db981352be236f79bb1db54b3"} Nov 24 13:29:14 crc kubenswrapper[4824]: I1124 13:29:14.158886 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7ab142b-a25e-4591-9672-6209bf35a3a8" containerID="3f33eafb596bc8f291db18a492241e384632f87cbf9a4d16ebe32ad43fcfba7e" exitCode=0 Nov 24 13:29:14 crc kubenswrapper[4824]: I1124 13:29:14.158974 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerDied","Data":"3f33eafb596bc8f291db18a492241e384632f87cbf9a4d16ebe32ad43fcfba7e"} Nov 24 13:29:15 crc kubenswrapper[4824]: I1124 13:29:15.123210 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-r5w24" Nov 24 13:29:15 crc kubenswrapper[4824]: I1124 13:29:15.167059 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerStarted","Data":"c73a5ac2bfa42cb0cc5df5bc8ab5b5e5065138c5e11ed8272181940dbe612c76"} Nov 24 13:29:15 crc kubenswrapper[4824]: I1124 13:29:15.167097 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerStarted","Data":"e55fd91ae742138121fc7dc1b0b299fbd992086aad84111bb9e96b1bb068f690"} Nov 24 13:29:15 crc kubenswrapper[4824]: I1124 13:29:15.167106 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerStarted","Data":"1bbe9ee38f3a7ad8cf297ff59ac44e3aacf8cf8dd5e46b76103160a225936303"} Nov 24 13:29:15 crc kubenswrapper[4824]: I1124 13:29:15.167113 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerStarted","Data":"ef2881d7188864cb1f9dda73937be95a7aa6346a3439b5cea8dbed234a5887af"} Nov 24 13:29:15 crc kubenswrapper[4824]: I1124 13:29:15.167122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerStarted","Data":"f8ce92e0d4968500bf26826aad0a085bbe0847da69246ee8a566374bf4cb2814"} Nov 24 13:29:16 crc kubenswrapper[4824]: I1124 13:29:16.187598 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dxtg8" event={"ID":"d7ab142b-a25e-4591-9672-6209bf35a3a8","Type":"ContainerStarted","Data":"c0d9c65f3effea1f49f91c99b635530b9249ac9d8fd55bb90ffc0a42cc389431"} Nov 24 13:29:16 crc kubenswrapper[4824]: I1124 13:29:16.189898 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:16 crc kubenswrapper[4824]: I1124 13:29:16.209162 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-dxtg8" podStartSLOduration=5.499658254 podStartE2EDuration="13.209144348s" podCreationTimestamp="2025-11-24 13:29:03 +0000 UTC" firstStartedPulling="2025-11-24 13:29:03.702996437 +0000 UTC m=+885.342535747" lastFinishedPulling="2025-11-24 13:29:11.412482531 +0000 UTC m=+893.052021841" observedRunningTime="2025-11-24 13:29:16.207190127 +0000 UTC m=+897.846729437" watchObservedRunningTime="2025-11-24 13:29:16.209144348 +0000 UTC m=+897.848683678" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.132744 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-w7klj"] Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.133859 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w7klj" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.136789 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.137123 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-5zrr9" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.138390 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.150639 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-w7klj"] Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.237894 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lt95\" (UniqueName: \"kubernetes.io/projected/99a3fc05-1519-4ed6-a906-261d2e6ffa77-kube-api-access-7lt95\") pod \"openstack-operator-index-w7klj\" (UID: \"99a3fc05-1519-4ed6-a906-261d2e6ffa77\") " pod="openstack-operators/openstack-operator-index-w7klj" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.339595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lt95\" (UniqueName: \"kubernetes.io/projected/99a3fc05-1519-4ed6-a906-261d2e6ffa77-kube-api-access-7lt95\") pod \"openstack-operator-index-w7klj\" (UID: \"99a3fc05-1519-4ed6-a906-261d2e6ffa77\") " pod="openstack-operators/openstack-operator-index-w7klj" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.368594 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lt95\" (UniqueName: \"kubernetes.io/projected/99a3fc05-1519-4ed6-a906-261d2e6ffa77-kube-api-access-7lt95\") pod \"openstack-operator-index-w7klj\" (UID: \"99a3fc05-1519-4ed6-a906-261d2e6ffa77\") " pod="openstack-operators/openstack-operator-index-w7klj" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.451623 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w7klj" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.503463 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.571164 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:18 crc kubenswrapper[4824]: I1124 13:29:18.692202 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-w7klj"] Nov 24 13:29:19 crc kubenswrapper[4824]: I1124 13:29:19.205920 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w7klj" event={"ID":"99a3fc05-1519-4ed6-a906-261d2e6ffa77","Type":"ContainerStarted","Data":"ee6614b0fee5659b16361afc66db9a1eba6433dae8849bd1286cfd5c4eb69f36"} Nov 24 13:29:21 crc kubenswrapper[4824]: I1124 13:29:21.511278 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-w7klj"] Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.129790 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hht55"] Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.130853 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hht55" Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.138117 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hht55"] Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.235637 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w7klj" event={"ID":"99a3fc05-1519-4ed6-a906-261d2e6ffa77","Type":"ContainerStarted","Data":"7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343"} Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.259090 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-w7klj" podStartSLOduration=1.847410137 podStartE2EDuration="4.259067254s" podCreationTimestamp="2025-11-24 13:29:18 +0000 UTC" firstStartedPulling="2025-11-24 13:29:18.702880103 +0000 UTC m=+900.342419423" lastFinishedPulling="2025-11-24 13:29:21.11453723 +0000 UTC m=+902.754076540" observedRunningTime="2025-11-24 13:29:22.254064093 +0000 UTC m=+903.893603413" watchObservedRunningTime="2025-11-24 13:29:22.259067254 +0000 UTC m=+903.898606584" Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.288490 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bh4v\" (UniqueName: \"kubernetes.io/projected/eaf96b33-49f7-4e4f-86e2-7a8a8b479bba-kube-api-access-8bh4v\") pod \"openstack-operator-index-hht55\" (UID: \"eaf96b33-49f7-4e4f-86e2-7a8a8b479bba\") " pod="openstack-operators/openstack-operator-index-hht55" Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.389622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bh4v\" (UniqueName: \"kubernetes.io/projected/eaf96b33-49f7-4e4f-86e2-7a8a8b479bba-kube-api-access-8bh4v\") pod \"openstack-operator-index-hht55\" (UID: \"eaf96b33-49f7-4e4f-86e2-7a8a8b479bba\") " pod="openstack-operators/openstack-operator-index-hht55" Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.410436 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bh4v\" (UniqueName: \"kubernetes.io/projected/eaf96b33-49f7-4e4f-86e2-7a8a8b479bba-kube-api-access-8bh4v\") pod \"openstack-operator-index-hht55\" (UID: \"eaf96b33-49f7-4e4f-86e2-7a8a8b479bba\") " pod="openstack-operators/openstack-operator-index-hht55" Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.459207 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hht55" Nov 24 13:29:22 crc kubenswrapper[4824]: I1124 13:29:22.840840 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hht55"] Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.243634 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hht55" event={"ID":"eaf96b33-49f7-4e4f-86e2-7a8a8b479bba","Type":"ContainerStarted","Data":"bca331548dd2d329438d0adf30b075c4514fa344ee9205c509a3864f555934f7"} Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.244057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hht55" event={"ID":"eaf96b33-49f7-4e4f-86e2-7a8a8b479bba","Type":"ContainerStarted","Data":"8ea28ea9b587acb19b082e867a866b8c3f2119d59b13d705b5c5ea49d8654014"} Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.243705 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-w7klj" podUID="99a3fc05-1519-4ed6-a906-261d2e6ffa77" containerName="registry-server" containerID="cri-o://7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343" gracePeriod=2 Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.265444 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hht55" podStartSLOduration=1.217007964 podStartE2EDuration="1.265417341s" podCreationTimestamp="2025-11-24 13:29:22 +0000 UTC" firstStartedPulling="2025-11-24 13:29:22.858368214 +0000 UTC m=+904.497907544" lastFinishedPulling="2025-11-24 13:29:22.906777611 +0000 UTC m=+904.546316921" observedRunningTime="2025-11-24 13:29:23.261841737 +0000 UTC m=+904.901381087" watchObservedRunningTime="2025-11-24 13:29:23.265417341 +0000 UTC m=+904.904956661" Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.520884 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2gp4q" Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.653495 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w7klj" Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.661304 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-m4d79" Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.807951 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lt95\" (UniqueName: \"kubernetes.io/projected/99a3fc05-1519-4ed6-a906-261d2e6ffa77-kube-api-access-7lt95\") pod \"99a3fc05-1519-4ed6-a906-261d2e6ffa77\" (UID: \"99a3fc05-1519-4ed6-a906-261d2e6ffa77\") " Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.819954 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99a3fc05-1519-4ed6-a906-261d2e6ffa77-kube-api-access-7lt95" (OuterVolumeSpecName: "kube-api-access-7lt95") pod "99a3fc05-1519-4ed6-a906-261d2e6ffa77" (UID: "99a3fc05-1519-4ed6-a906-261d2e6ffa77"). InnerVolumeSpecName "kube-api-access-7lt95". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:29:23 crc kubenswrapper[4824]: I1124 13:29:23.910109 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lt95\" (UniqueName: \"kubernetes.io/projected/99a3fc05-1519-4ed6-a906-261d2e6ffa77-kube-api-access-7lt95\") on node \"crc\" DevicePath \"\"" Nov 24 13:29:24 crc kubenswrapper[4824]: I1124 13:29:24.250767 4824 generic.go:334] "Generic (PLEG): container finished" podID="99a3fc05-1519-4ed6-a906-261d2e6ffa77" containerID="7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343" exitCode=0 Nov 24 13:29:24 crc kubenswrapper[4824]: I1124 13:29:24.250868 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w7klj" event={"ID":"99a3fc05-1519-4ed6-a906-261d2e6ffa77","Type":"ContainerDied","Data":"7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343"} Nov 24 13:29:24 crc kubenswrapper[4824]: I1124 13:29:24.250916 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w7klj" event={"ID":"99a3fc05-1519-4ed6-a906-261d2e6ffa77","Type":"ContainerDied","Data":"ee6614b0fee5659b16361afc66db9a1eba6433dae8849bd1286cfd5c4eb69f36"} Nov 24 13:29:24 crc kubenswrapper[4824]: I1124 13:29:24.250947 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w7klj" Nov 24 13:29:24 crc kubenswrapper[4824]: I1124 13:29:24.250961 4824 scope.go:117] "RemoveContainer" containerID="7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343" Nov 24 13:29:24 crc kubenswrapper[4824]: I1124 13:29:24.277038 4824 scope.go:117] "RemoveContainer" containerID="7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343" Nov 24 13:29:24 crc kubenswrapper[4824]: E1124 13:29:24.278400 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343\": container with ID starting with 7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343 not found: ID does not exist" containerID="7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343" Nov 24 13:29:24 crc kubenswrapper[4824]: I1124 13:29:24.278462 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343"} err="failed to get container status \"7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343\": rpc error: code = NotFound desc = could not find container \"7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343\": container with ID starting with 7b34ad6fee08bb3560c03f0a285c9fdc890d6a3cc8cf4bd3675def7c7cd94343 not found: ID does not exist" Nov 24 13:29:24 crc kubenswrapper[4824]: I1124 13:29:24.282311 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-w7klj"] Nov 24 13:29:24 crc kubenswrapper[4824]: I1124 13:29:24.286517 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-w7klj"] Nov 24 13:29:25 crc kubenswrapper[4824]: I1124 13:29:25.025175 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99a3fc05-1519-4ed6-a906-261d2e6ffa77" path="/var/lib/kubelet/pods/99a3fc05-1519-4ed6-a906-261d2e6ffa77/volumes" Nov 24 13:29:32 crc kubenswrapper[4824]: I1124 13:29:32.459854 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hht55" Nov 24 13:29:32 crc kubenswrapper[4824]: I1124 13:29:32.460747 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hht55" Nov 24 13:29:32 crc kubenswrapper[4824]: I1124 13:29:32.496959 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hht55" Nov 24 13:29:33 crc kubenswrapper[4824]: I1124 13:29:33.337027 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hht55" Nov 24 13:29:33 crc kubenswrapper[4824]: I1124 13:29:33.505674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-dxtg8" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.593741 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z"] Nov 24 13:29:39 crc kubenswrapper[4824]: E1124 13:29:39.594622 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99a3fc05-1519-4ed6-a906-261d2e6ffa77" containerName="registry-server" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.594652 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="99a3fc05-1519-4ed6-a906-261d2e6ffa77" containerName="registry-server" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.594943 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="99a3fc05-1519-4ed6-a906-261d2e6ffa77" containerName="registry-server" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.596923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.601152 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-xl6cx" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.609657 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z"] Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.736386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-bundle\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.736603 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-util\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.736658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsrn8\" (UniqueName: \"kubernetes.io/projected/d720b626-1d35-4d40-95be-dbd95dcafc27-kube-api-access-fsrn8\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.837723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-util\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.837791 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsrn8\" (UniqueName: \"kubernetes.io/projected/d720b626-1d35-4d40-95be-dbd95dcafc27-kube-api-access-fsrn8\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.837876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-bundle\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.838272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-util\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.838304 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-bundle\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.857739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsrn8\" (UniqueName: \"kubernetes.io/projected/d720b626-1d35-4d40-95be-dbd95dcafc27-kube-api-access-fsrn8\") pod \"f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:39 crc kubenswrapper[4824]: I1124 13:29:39.922537 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:40 crc kubenswrapper[4824]: I1124 13:29:40.375999 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z"] Nov 24 13:29:41 crc kubenswrapper[4824]: I1124 13:29:41.370111 4824 generic.go:334] "Generic (PLEG): container finished" podID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerID="cf50d9e27bc486546d7d9cae159ceca03fe19f01ced05146f8163589f86a2143" exitCode=0 Nov 24 13:29:41 crc kubenswrapper[4824]: I1124 13:29:41.370166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" event={"ID":"d720b626-1d35-4d40-95be-dbd95dcafc27","Type":"ContainerDied","Data":"cf50d9e27bc486546d7d9cae159ceca03fe19f01ced05146f8163589f86a2143"} Nov 24 13:29:41 crc kubenswrapper[4824]: I1124 13:29:41.370266 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" event={"ID":"d720b626-1d35-4d40-95be-dbd95dcafc27","Type":"ContainerStarted","Data":"392065a2c46755a494f6bc64e92913dbde83a7f20a6f791a682bc856490aae52"} Nov 24 13:29:42 crc kubenswrapper[4824]: I1124 13:29:42.377921 4824 generic.go:334] "Generic (PLEG): container finished" podID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerID="628ac0deb6fd68d34f4867372154d0ad087948595f8562892baa6348c3cdb4f3" exitCode=0 Nov 24 13:29:42 crc kubenswrapper[4824]: I1124 13:29:42.377963 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" event={"ID":"d720b626-1d35-4d40-95be-dbd95dcafc27","Type":"ContainerDied","Data":"628ac0deb6fd68d34f4867372154d0ad087948595f8562892baa6348c3cdb4f3"} Nov 24 13:29:43 crc kubenswrapper[4824]: I1124 13:29:43.387561 4824 generic.go:334] "Generic (PLEG): container finished" podID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerID="beffccd217f0860b6846b79dc6f64c675d892dac5b6cd4f43bb57fe895ba675b" exitCode=0 Nov 24 13:29:43 crc kubenswrapper[4824]: I1124 13:29:43.387640 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" event={"ID":"d720b626-1d35-4d40-95be-dbd95dcafc27","Type":"ContainerDied","Data":"beffccd217f0860b6846b79dc6f64c675d892dac5b6cd4f43bb57fe895ba675b"} Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.625036 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.708738 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-util\") pod \"d720b626-1d35-4d40-95be-dbd95dcafc27\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.708872 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-bundle\") pod \"d720b626-1d35-4d40-95be-dbd95dcafc27\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.709583 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-bundle" (OuterVolumeSpecName: "bundle") pod "d720b626-1d35-4d40-95be-dbd95dcafc27" (UID: "d720b626-1d35-4d40-95be-dbd95dcafc27"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.709775 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsrn8\" (UniqueName: \"kubernetes.io/projected/d720b626-1d35-4d40-95be-dbd95dcafc27-kube-api-access-fsrn8\") pod \"d720b626-1d35-4d40-95be-dbd95dcafc27\" (UID: \"d720b626-1d35-4d40-95be-dbd95dcafc27\") " Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.710793 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.715597 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d720b626-1d35-4d40-95be-dbd95dcafc27-kube-api-access-fsrn8" (OuterVolumeSpecName: "kube-api-access-fsrn8") pod "d720b626-1d35-4d40-95be-dbd95dcafc27" (UID: "d720b626-1d35-4d40-95be-dbd95dcafc27"). InnerVolumeSpecName "kube-api-access-fsrn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.728032 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-util" (OuterVolumeSpecName: "util") pod "d720b626-1d35-4d40-95be-dbd95dcafc27" (UID: "d720b626-1d35-4d40-95be-dbd95dcafc27"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.812287 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsrn8\" (UniqueName: \"kubernetes.io/projected/d720b626-1d35-4d40-95be-dbd95dcafc27-kube-api-access-fsrn8\") on node \"crc\" DevicePath \"\"" Nov 24 13:29:44 crc kubenswrapper[4824]: I1124 13:29:44.812316 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d720b626-1d35-4d40-95be-dbd95dcafc27-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:29:45 crc kubenswrapper[4824]: I1124 13:29:45.401632 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" event={"ID":"d720b626-1d35-4d40-95be-dbd95dcafc27","Type":"ContainerDied","Data":"392065a2c46755a494f6bc64e92913dbde83a7f20a6f791a682bc856490aae52"} Nov 24 13:29:45 crc kubenswrapper[4824]: I1124 13:29:45.401684 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="392065a2c46755a494f6bc64e92913dbde83a7f20a6f791a682bc856490aae52" Nov 24 13:29:45 crc kubenswrapper[4824]: I1124 13:29:45.401747 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z" Nov 24 13:29:49 crc kubenswrapper[4824]: I1124 13:29:49.867682 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq"] Nov 24 13:29:49 crc kubenswrapper[4824]: E1124 13:29:49.868110 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerName="extract" Nov 24 13:29:49 crc kubenswrapper[4824]: I1124 13:29:49.868120 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerName="extract" Nov 24 13:29:49 crc kubenswrapper[4824]: E1124 13:29:49.868132 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerName="pull" Nov 24 13:29:49 crc kubenswrapper[4824]: I1124 13:29:49.868137 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerName="pull" Nov 24 13:29:49 crc kubenswrapper[4824]: E1124 13:29:49.868154 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerName="util" Nov 24 13:29:49 crc kubenswrapper[4824]: I1124 13:29:49.868161 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerName="util" Nov 24 13:29:49 crc kubenswrapper[4824]: I1124 13:29:49.868269 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d720b626-1d35-4d40-95be-dbd95dcafc27" containerName="extract" Nov 24 13:29:49 crc kubenswrapper[4824]: I1124 13:29:49.868848 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" Nov 24 13:29:49 crc kubenswrapper[4824]: I1124 13:29:49.872552 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-vqhkm" Nov 24 13:29:49 crc kubenswrapper[4824]: I1124 13:29:49.899082 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq"] Nov 24 13:29:50 crc kubenswrapper[4824]: I1124 13:29:50.002966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzdm8\" (UniqueName: \"kubernetes.io/projected/af6dc481-60ca-48ae-a75c-edec34270b7f-kube-api-access-zzdm8\") pod \"openstack-operator-controller-operator-866b74d8b7-d68dq\" (UID: \"af6dc481-60ca-48ae-a75c-edec34270b7f\") " pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" Nov 24 13:29:50 crc kubenswrapper[4824]: I1124 13:29:50.104225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzdm8\" (UniqueName: \"kubernetes.io/projected/af6dc481-60ca-48ae-a75c-edec34270b7f-kube-api-access-zzdm8\") pod \"openstack-operator-controller-operator-866b74d8b7-d68dq\" (UID: \"af6dc481-60ca-48ae-a75c-edec34270b7f\") " pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" Nov 24 13:29:50 crc kubenswrapper[4824]: I1124 13:29:50.142406 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzdm8\" (UniqueName: \"kubernetes.io/projected/af6dc481-60ca-48ae-a75c-edec34270b7f-kube-api-access-zzdm8\") pod \"openstack-operator-controller-operator-866b74d8b7-d68dq\" (UID: \"af6dc481-60ca-48ae-a75c-edec34270b7f\") " pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" Nov 24 13:29:50 crc kubenswrapper[4824]: I1124 13:29:50.183027 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" Nov 24 13:29:50 crc kubenswrapper[4824]: I1124 13:29:50.771779 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq"] Nov 24 13:29:51 crc kubenswrapper[4824]: I1124 13:29:51.438610 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" event={"ID":"af6dc481-60ca-48ae-a75c-edec34270b7f","Type":"ContainerStarted","Data":"06e3bff2c07526c8e3c6ec5e9b64631466016ae533dc7413b1d9c98ce88e7249"} Nov 24 13:29:57 crc kubenswrapper[4824]: I1124 13:29:57.493852 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" event={"ID":"af6dc481-60ca-48ae-a75c-edec34270b7f","Type":"ContainerStarted","Data":"988ca742748a1e0ca2b0eefaa8f0c7c7564dcc81bf191065f65dde02766a2de4"} Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.120613 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c"] Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.121740 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.123590 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.126370 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.131312 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c"] Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.247925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/007fcfe8-9560-471e-9cd0-c4821bffa2d4-secret-volume\") pod \"collect-profiles-29399850-tmf7c\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.248015 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/007fcfe8-9560-471e-9cd0-c4821bffa2d4-config-volume\") pod \"collect-profiles-29399850-tmf7c\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.248056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2thgn\" (UniqueName: \"kubernetes.io/projected/007fcfe8-9560-471e-9cd0-c4821bffa2d4-kube-api-access-2thgn\") pod \"collect-profiles-29399850-tmf7c\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.348951 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2thgn\" (UniqueName: \"kubernetes.io/projected/007fcfe8-9560-471e-9cd0-c4821bffa2d4-kube-api-access-2thgn\") pod \"collect-profiles-29399850-tmf7c\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.349028 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/007fcfe8-9560-471e-9cd0-c4821bffa2d4-secret-volume\") pod \"collect-profiles-29399850-tmf7c\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.349062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/007fcfe8-9560-471e-9cd0-c4821bffa2d4-config-volume\") pod \"collect-profiles-29399850-tmf7c\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.350573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/007fcfe8-9560-471e-9cd0-c4821bffa2d4-config-volume\") pod \"collect-profiles-29399850-tmf7c\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.363684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/007fcfe8-9560-471e-9cd0-c4821bffa2d4-secret-volume\") pod \"collect-profiles-29399850-tmf7c\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.368759 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2thgn\" (UniqueName: \"kubernetes.io/projected/007fcfe8-9560-471e-9cd0-c4821bffa2d4-kube-api-access-2thgn\") pod \"collect-profiles-29399850-tmf7c\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:00 crc kubenswrapper[4824]: I1124 13:30:00.437770 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:03 crc kubenswrapper[4824]: I1124 13:30:03.646414 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c"] Nov 24 13:30:03 crc kubenswrapper[4824]: W1124 13:30:03.655059 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod007fcfe8_9560_471e_9cd0_c4821bffa2d4.slice/crio-69403fdf532d2364fd9b1c155992d3001d20f850c18b59d3109b889dae8d53a5 WatchSource:0}: Error finding container 69403fdf532d2364fd9b1c155992d3001d20f850c18b59d3109b889dae8d53a5: Status 404 returned error can't find the container with id 69403fdf532d2364fd9b1c155992d3001d20f850c18b59d3109b889dae8d53a5 Nov 24 13:30:04 crc kubenswrapper[4824]: I1124 13:30:04.551256 4824 generic.go:334] "Generic (PLEG): container finished" podID="007fcfe8-9560-471e-9cd0-c4821bffa2d4" containerID="4a7d9bfbb6bc4d3d2ed940525a1df1d99010a102ca5c61ca90d0d5c7d11c683e" exitCode=0 Nov 24 13:30:04 crc kubenswrapper[4824]: I1124 13:30:04.551309 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" event={"ID":"007fcfe8-9560-471e-9cd0-c4821bffa2d4","Type":"ContainerDied","Data":"4a7d9bfbb6bc4d3d2ed940525a1df1d99010a102ca5c61ca90d0d5c7d11c683e"} Nov 24 13:30:04 crc kubenswrapper[4824]: I1124 13:30:04.551692 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" event={"ID":"007fcfe8-9560-471e-9cd0-c4821bffa2d4","Type":"ContainerStarted","Data":"69403fdf532d2364fd9b1c155992d3001d20f850c18b59d3109b889dae8d53a5"} Nov 24 13:30:05 crc kubenswrapper[4824]: I1124 13:30:05.560541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" event={"ID":"af6dc481-60ca-48ae-a75c-edec34270b7f","Type":"ContainerStarted","Data":"73d69b7aedcfe0b0401b9748b0f1f06414a5537e9cdb2520cf2344a7ee322a7c"} Nov 24 13:30:05 crc kubenswrapper[4824]: I1124 13:30:05.600881 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" podStartSLOduration=2.884592742 podStartE2EDuration="16.600855224s" podCreationTimestamp="2025-11-24 13:29:49 +0000 UTC" firstStartedPulling="2025-11-24 13:29:50.784744477 +0000 UTC m=+932.424283787" lastFinishedPulling="2025-11-24 13:30:04.501006959 +0000 UTC m=+946.140546269" observedRunningTime="2025-11-24 13:30:05.596770329 +0000 UTC m=+947.236309669" watchObservedRunningTime="2025-11-24 13:30:05.600855224 +0000 UTC m=+947.240394554" Nov 24 13:30:05 crc kubenswrapper[4824]: I1124 13:30:05.817749 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:05 crc kubenswrapper[4824]: I1124 13:30:05.921709 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2thgn\" (UniqueName: \"kubernetes.io/projected/007fcfe8-9560-471e-9cd0-c4821bffa2d4-kube-api-access-2thgn\") pod \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " Nov 24 13:30:05 crc kubenswrapper[4824]: I1124 13:30:05.921820 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/007fcfe8-9560-471e-9cd0-c4821bffa2d4-secret-volume\") pod \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " Nov 24 13:30:05 crc kubenswrapper[4824]: I1124 13:30:05.921882 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/007fcfe8-9560-471e-9cd0-c4821bffa2d4-config-volume\") pod \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\" (UID: \"007fcfe8-9560-471e-9cd0-c4821bffa2d4\") " Nov 24 13:30:05 crc kubenswrapper[4824]: I1124 13:30:05.922896 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/007fcfe8-9560-471e-9cd0-c4821bffa2d4-config-volume" (OuterVolumeSpecName: "config-volume") pod "007fcfe8-9560-471e-9cd0-c4821bffa2d4" (UID: "007fcfe8-9560-471e-9cd0-c4821bffa2d4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:30:05 crc kubenswrapper[4824]: I1124 13:30:05.927028 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/007fcfe8-9560-471e-9cd0-c4821bffa2d4-kube-api-access-2thgn" (OuterVolumeSpecName: "kube-api-access-2thgn") pod "007fcfe8-9560-471e-9cd0-c4821bffa2d4" (UID: "007fcfe8-9560-471e-9cd0-c4821bffa2d4"). InnerVolumeSpecName "kube-api-access-2thgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:30:05 crc kubenswrapper[4824]: I1124 13:30:05.928686 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/007fcfe8-9560-471e-9cd0-c4821bffa2d4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "007fcfe8-9560-471e-9cd0-c4821bffa2d4" (UID: "007fcfe8-9560-471e-9cd0-c4821bffa2d4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:30:06 crc kubenswrapper[4824]: I1124 13:30:06.022842 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/007fcfe8-9560-471e-9cd0-c4821bffa2d4-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:30:06 crc kubenswrapper[4824]: I1124 13:30:06.022894 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/007fcfe8-9560-471e-9cd0-c4821bffa2d4-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:30:06 crc kubenswrapper[4824]: I1124 13:30:06.022911 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2thgn\" (UniqueName: \"kubernetes.io/projected/007fcfe8-9560-471e-9cd0-c4821bffa2d4-kube-api-access-2thgn\") on node \"crc\" DevicePath \"\"" Nov 24 13:30:06 crc kubenswrapper[4824]: I1124 13:30:06.576767 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" event={"ID":"007fcfe8-9560-471e-9cd0-c4821bffa2d4","Type":"ContainerDied","Data":"69403fdf532d2364fd9b1c155992d3001d20f850c18b59d3109b889dae8d53a5"} Nov 24 13:30:06 crc kubenswrapper[4824]: I1124 13:30:06.576852 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69403fdf532d2364fd9b1c155992d3001d20f850c18b59d3109b889dae8d53a5" Nov 24 13:30:06 crc kubenswrapper[4824]: I1124 13:30:06.576852 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c" Nov 24 13:30:06 crc kubenswrapper[4824]: I1124 13:30:06.577351 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" Nov 24 13:30:06 crc kubenswrapper[4824]: I1124 13:30:06.580846 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-866b74d8b7-d68dq" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.364053 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg"] Nov 24 13:30:23 crc kubenswrapper[4824]: E1124 13:30:23.366002 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="007fcfe8-9560-471e-9cd0-c4821bffa2d4" containerName="collect-profiles" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.366133 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="007fcfe8-9560-471e-9cd0-c4821bffa2d4" containerName="collect-profiles" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.366355 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="007fcfe8-9560-471e-9cd0-c4821bffa2d4" containerName="collect-profiles" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.367284 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.369833 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-smbw2" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.380852 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.386542 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.391331 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-rxrxv" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.401181 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.436742 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.440329 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.441435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.444751 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-zvzrb" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.450425 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.452132 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.465573 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-jxwlz" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.476747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98vgw\" (UniqueName: \"kubernetes.io/projected/dc25960c-da62-4fe3-97c1-cf8526bc9c1d-kube-api-access-98vgw\") pod \"barbican-operator-controller-manager-75fb479bcc-747cg\" (UID: \"dc25960c-da62-4fe3-97c1-cf8526bc9c1d\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.476787 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfm4c\" (UniqueName: \"kubernetes.io/projected/88cfde0a-21e8-45ca-a7b3-43699c2ff345-kube-api-access-xfm4c\") pod \"cinder-operator-controller-manager-6498cbf48f-h5mx8\" (UID: \"88cfde0a-21e8-45ca-a7b3-43699c2ff345\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.478880 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.493104 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.505312 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.506290 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.515418 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-lb4cz" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.525896 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.549237 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.550131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.553442 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-ws8bm" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.579428 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlcrm\" (UniqueName: \"kubernetes.io/projected/3b95ed18-ec44-4712-9f01-adeef55deb75-kube-api-access-wlcrm\") pod \"designate-operator-controller-manager-767ccfd65f-gw957\" (UID: \"3b95ed18-ec44-4712-9f01-adeef55deb75\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.579473 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwflw\" (UniqueName: \"kubernetes.io/projected/0cac78b6-e909-4616-bce8-d3ce8ac90452-kube-api-access-wwflw\") pod \"heat-operator-controller-manager-56f54d6746-s2k8r\" (UID: \"0cac78b6-e909-4616-bce8-d3ce8ac90452\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.579509 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98vgw\" (UniqueName: \"kubernetes.io/projected/dc25960c-da62-4fe3-97c1-cf8526bc9c1d-kube-api-access-98vgw\") pod \"barbican-operator-controller-manager-75fb479bcc-747cg\" (UID: \"dc25960c-da62-4fe3-97c1-cf8526bc9c1d\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.579529 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n9q2\" (UniqueName: \"kubernetes.io/projected/9c973000-24b9-4087-a3eb-c5f087a50ae3-kube-api-access-7n9q2\") pod \"glance-operator-controller-manager-7969689c84-gvhjd\" (UID: \"9c973000-24b9-4087-a3eb-c5f087a50ae3\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.579547 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfm4c\" (UniqueName: \"kubernetes.io/projected/88cfde0a-21e8-45ca-a7b3-43699c2ff345-kube-api-access-xfm4c\") pod \"cinder-operator-controller-manager-6498cbf48f-h5mx8\" (UID: \"88cfde0a-21e8-45ca-a7b3-43699c2ff345\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.583990 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.585573 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.589036 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.589056 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4drwk" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.593577 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.608221 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.612760 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.613727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.625672 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-fkwwl" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.643710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98vgw\" (UniqueName: \"kubernetes.io/projected/dc25960c-da62-4fe3-97c1-cf8526bc9c1d-kube-api-access-98vgw\") pod \"barbican-operator-controller-manager-75fb479bcc-747cg\" (UID: \"dc25960c-da62-4fe3-97c1-cf8526bc9c1d\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.655479 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.662328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfm4c\" (UniqueName: \"kubernetes.io/projected/88cfde0a-21e8-45ca-a7b3-43699c2ff345-kube-api-access-xfm4c\") pod \"cinder-operator-controller-manager-6498cbf48f-h5mx8\" (UID: \"88cfde0a-21e8-45ca-a7b3-43699c2ff345\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.688104 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.688362 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78gsx\" (UniqueName: \"kubernetes.io/projected/5f502855-0eb1-455d-8e3b-4af6ac353aa9-kube-api-access-78gsx\") pod \"horizon-operator-controller-manager-598f69df5d-cdnsz\" (UID: \"5f502855-0eb1-455d-8e3b-4af6ac353aa9\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.688432 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlcrm\" (UniqueName: \"kubernetes.io/projected/3b95ed18-ec44-4712-9f01-adeef55deb75-kube-api-access-wlcrm\") pod \"designate-operator-controller-manager-767ccfd65f-gw957\" (UID: \"3b95ed18-ec44-4712-9f01-adeef55deb75\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.688459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwflw\" (UniqueName: \"kubernetes.io/projected/0cac78b6-e909-4616-bce8-d3ce8ac90452-kube-api-access-wwflw\") pod \"heat-operator-controller-manager-56f54d6746-s2k8r\" (UID: \"0cac78b6-e909-4616-bce8-d3ce8ac90452\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.688490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n9q2\" (UniqueName: \"kubernetes.io/projected/9c973000-24b9-4087-a3eb-c5f087a50ae3-kube-api-access-7n9q2\") pod \"glance-operator-controller-manager-7969689c84-gvhjd\" (UID: \"9c973000-24b9-4087-a3eb-c5f087a50ae3\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.688509 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec73cd81-275a-44e0-93dc-6b20123a88d3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-rn29q\" (UID: \"ec73cd81-275a-44e0-93dc-6b20123a88d3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.688536 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pvqj\" (UniqueName: \"kubernetes.io/projected/ec73cd81-275a-44e0-93dc-6b20123a88d3-kube-api-access-9pvqj\") pod \"infra-operator-controller-manager-6dd8864d7c-rn29q\" (UID: \"ec73cd81-275a-44e0-93dc-6b20123a88d3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.694856 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.696012 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.710214 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.710318 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-ffvmd" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.716971 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.718466 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.731884 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-dl5rf" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.733591 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.743481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n9q2\" (UniqueName: \"kubernetes.io/projected/9c973000-24b9-4087-a3eb-c5f087a50ae3-kube-api-access-7n9q2\") pod \"glance-operator-controller-manager-7969689c84-gvhjd\" (UID: \"9c973000-24b9-4087-a3eb-c5f087a50ae3\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.756167 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.757370 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.760084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-77pz6" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.761286 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlcrm\" (UniqueName: \"kubernetes.io/projected/3b95ed18-ec44-4712-9f01-adeef55deb75-kube-api-access-wlcrm\") pod \"designate-operator-controller-manager-767ccfd65f-gw957\" (UID: \"3b95ed18-ec44-4712-9f01-adeef55deb75\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.776582 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.777739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwflw\" (UniqueName: \"kubernetes.io/projected/0cac78b6-e909-4616-bce8-d3ce8ac90452-kube-api-access-wwflw\") pod \"heat-operator-controller-manager-56f54d6746-s2k8r\" (UID: \"0cac78b6-e909-4616-bce8-d3ce8ac90452\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.778396 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.797136 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.806083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec73cd81-275a-44e0-93dc-6b20123a88d3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-rn29q\" (UID: \"ec73cd81-275a-44e0-93dc-6b20123a88d3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.806159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmqkn\" (UniqueName: \"kubernetes.io/projected/05a9eb12-a2a1-4fbe-901b-cb817e4713b4-kube-api-access-wmqkn\") pod \"ironic-operator-controller-manager-99b499f4-cmmdj\" (UID: \"05a9eb12-a2a1-4fbe-901b-cb817e4713b4\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.806200 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pvqj\" (UniqueName: \"kubernetes.io/projected/ec73cd81-275a-44e0-93dc-6b20123a88d3-kube-api-access-9pvqj\") pod \"infra-operator-controller-manager-6dd8864d7c-rn29q\" (UID: \"ec73cd81-275a-44e0-93dc-6b20123a88d3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.806232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78gsx\" (UniqueName: \"kubernetes.io/projected/5f502855-0eb1-455d-8e3b-4af6ac353aa9-kube-api-access-78gsx\") pod \"horizon-operator-controller-manager-598f69df5d-cdnsz\" (UID: \"5f502855-0eb1-455d-8e3b-4af6ac353aa9\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.806267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxnt6\" (UniqueName: \"kubernetes.io/projected/65d269bf-f6af-4239-8e51-bf44fe7830d1-kube-api-access-bxnt6\") pod \"keystone-operator-controller-manager-7454b96578-v6dbs\" (UID: \"65d269bf-f6af-4239-8e51-bf44fe7830d1\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" Nov 24 13:30:23 crc kubenswrapper[4824]: E1124 13:30:23.806453 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 13:30:23 crc kubenswrapper[4824]: E1124 13:30:23.806499 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec73cd81-275a-44e0-93dc-6b20123a88d3-cert podName:ec73cd81-275a-44e0-93dc-6b20123a88d3 nodeName:}" failed. No retries permitted until 2025-11-24 13:30:24.306480204 +0000 UTC m=+965.946019514 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ec73cd81-275a-44e0-93dc-6b20123a88d3-cert") pod "infra-operator-controller-manager-6dd8864d7c-rn29q" (UID: "ec73cd81-275a-44e0-93dc-6b20123a88d3") : secret "infra-operator-webhook-server-cert" not found Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.813318 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.817993 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.828661 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.843259 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-qv96r" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.848645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78gsx\" (UniqueName: \"kubernetes.io/projected/5f502855-0eb1-455d-8e3b-4af6ac353aa9-kube-api-access-78gsx\") pod \"horizon-operator-controller-manager-598f69df5d-cdnsz\" (UID: \"5f502855-0eb1-455d-8e3b-4af6ac353aa9\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.850469 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.856089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.869194 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.872099 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pvqj\" (UniqueName: \"kubernetes.io/projected/ec73cd81-275a-44e0-93dc-6b20123a88d3-kube-api-access-9pvqj\") pod \"infra-operator-controller-manager-6dd8864d7c-rn29q\" (UID: \"ec73cd81-275a-44e0-93dc-6b20123a88d3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.872512 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.873926 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.885267 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx"] Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.887539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-whmfd" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.887986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.909593 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-dgwww" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.964144 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stgrc\" (UniqueName: \"kubernetes.io/projected/218ca72b-e49f-4830-8be8-9b6e85956009-kube-api-access-stgrc\") pod \"octavia-operator-controller-manager-54cfbf4c7d-bg2m4\" (UID: \"218ca72b-e49f-4830-8be8-9b6e85956009\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.964407 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcm5n\" (UniqueName: \"kubernetes.io/projected/3b3da19f-12b1-413d-b6aa-38bc498172c1-kube-api-access-wcm5n\") pod \"manila-operator-controller-manager-58f887965d-xdrwd\" (UID: \"3b3da19f-12b1-413d-b6aa-38bc498172c1\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.965876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv29s\" (UniqueName: \"kubernetes.io/projected/9971e72a-bb18-4951-a752-ee96f1725801-kube-api-access-hv29s\") pod \"mariadb-operator-controller-manager-54b5986bb8-mzs9m\" (UID: \"9971e72a-bb18-4951-a752-ee96f1725801\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.965982 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmqkn\" (UniqueName: \"kubernetes.io/projected/05a9eb12-a2a1-4fbe-901b-cb817e4713b4-kube-api-access-wmqkn\") pod \"ironic-operator-controller-manager-99b499f4-cmmdj\" (UID: \"05a9eb12-a2a1-4fbe-901b-cb817e4713b4\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.966127 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxnt6\" (UniqueName: \"kubernetes.io/projected/65d269bf-f6af-4239-8e51-bf44fe7830d1-kube-api-access-bxnt6\") pod \"keystone-operator-controller-manager-7454b96578-v6dbs\" (UID: \"65d269bf-f6af-4239-8e51-bf44fe7830d1\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" Nov 24 13:30:23 crc kubenswrapper[4824]: I1124 13:30:23.990103 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.023760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmqkn\" (UniqueName: \"kubernetes.io/projected/05a9eb12-a2a1-4fbe-901b-cb817e4713b4-kube-api-access-wmqkn\") pod \"ironic-operator-controller-manager-99b499f4-cmmdj\" (UID: \"05a9eb12-a2a1-4fbe-901b-cb817e4713b4\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.062033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxnt6\" (UniqueName: \"kubernetes.io/projected/65d269bf-f6af-4239-8e51-bf44fe7830d1-kube-api-access-bxnt6\") pod \"keystone-operator-controller-manager-7454b96578-v6dbs\" (UID: \"65d269bf-f6af-4239-8e51-bf44fe7830d1\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.065847 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.070644 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stgrc\" (UniqueName: \"kubernetes.io/projected/218ca72b-e49f-4830-8be8-9b6e85956009-kube-api-access-stgrc\") pod \"octavia-operator-controller-manager-54cfbf4c7d-bg2m4\" (UID: \"218ca72b-e49f-4830-8be8-9b6e85956009\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.070694 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjgj6\" (UniqueName: \"kubernetes.io/projected/4cca949a-2334-4d41-bb9e-13e6bd89d7a8-kube-api-access-fjgj6\") pod \"nova-operator-controller-manager-cfbb9c588-qzxtx\" (UID: \"4cca949a-2334-4d41-bb9e-13e6bd89d7a8\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.070715 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcm5n\" (UniqueName: \"kubernetes.io/projected/3b3da19f-12b1-413d-b6aa-38bc498172c1-kube-api-access-wcm5n\") pod \"manila-operator-controller-manager-58f887965d-xdrwd\" (UID: \"3b3da19f-12b1-413d-b6aa-38bc498172c1\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.070749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv29s\" (UniqueName: \"kubernetes.io/projected/9971e72a-bb18-4951-a752-ee96f1725801-kube-api-access-hv29s\") pod \"mariadb-operator-controller-manager-54b5986bb8-mzs9m\" (UID: \"9971e72a-bb18-4951-a752-ee96f1725801\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.070798 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qppbf\" (UniqueName: \"kubernetes.io/projected/50437b59-29ab-4e49-a751-80a83d9e2aa0-kube-api-access-qppbf\") pod \"neutron-operator-controller-manager-78bd47f458-zncj6\" (UID: \"50437b59-29ab-4e49-a751-80a83d9e2aa0\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.092114 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.113539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv29s\" (UniqueName: \"kubernetes.io/projected/9971e72a-bb18-4951-a752-ee96f1725801-kube-api-access-hv29s\") pod \"mariadb-operator-controller-manager-54b5986bb8-mzs9m\" (UID: \"9971e72a-bb18-4951-a752-ee96f1725801\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.116840 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stgrc\" (UniqueName: \"kubernetes.io/projected/218ca72b-e49f-4830-8be8-9b6e85956009-kube-api-access-stgrc\") pod \"octavia-operator-controller-manager-54cfbf4c7d-bg2m4\" (UID: \"218ca72b-e49f-4830-8be8-9b6e85956009\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.117669 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.128149 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcm5n\" (UniqueName: \"kubernetes.io/projected/3b3da19f-12b1-413d-b6aa-38bc498172c1-kube-api-access-wcm5n\") pod \"manila-operator-controller-manager-58f887965d-xdrwd\" (UID: \"3b3da19f-12b1-413d-b6aa-38bc498172c1\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.132311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.139308 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.145236 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.145571 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-t587t" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.149298 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.150515 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.153982 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-vz2n9" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.171615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qppbf\" (UniqueName: \"kubernetes.io/projected/50437b59-29ab-4e49-a751-80a83d9e2aa0-kube-api-access-qppbf\") pod \"neutron-operator-controller-manager-78bd47f458-zncj6\" (UID: \"50437b59-29ab-4e49-a751-80a83d9e2aa0\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.171698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjgj6\" (UniqueName: \"kubernetes.io/projected/4cca949a-2334-4d41-bb9e-13e6bd89d7a8-kube-api-access-fjgj6\") pod \"nova-operator-controller-manager-cfbb9c588-qzxtx\" (UID: \"4cca949a-2334-4d41-bb9e-13e6bd89d7a8\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.181530 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.238531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qppbf\" (UniqueName: \"kubernetes.io/projected/50437b59-29ab-4e49-a751-80a83d9e2aa0-kube-api-access-qppbf\") pod \"neutron-operator-controller-manager-78bd47f458-zncj6\" (UID: \"50437b59-29ab-4e49-a751-80a83d9e2aa0\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.247129 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.248533 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.251784 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-sp2d8" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.259166 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.273476 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99379416-bd58-4014-a297-7b14b8ff4aeb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp\" (UID: \"99379416-bd58-4014-a297-7b14b8ff4aeb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.273871 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz7cd\" (UniqueName: \"kubernetes.io/projected/99379416-bd58-4014-a297-7b14b8ff4aeb-kube-api-access-tz7cd\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp\" (UID: \"99379416-bd58-4014-a297-7b14b8ff4aeb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.274205 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kln7p\" (UniqueName: \"kubernetes.io/projected/926db806-d562-44f8-9ac9-71f337dbb202-kube-api-access-kln7p\") pod \"ovn-operator-controller-manager-54fc5f65b7-2jjxl\" (UID: \"926db806-d562-44f8-9ac9-71f337dbb202\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.275318 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.285127 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.286126 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.288397 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-wcfwx" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.290432 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjgj6\" (UniqueName: \"kubernetes.io/projected/4cca949a-2334-4d41-bb9e-13e6bd89d7a8-kube-api-access-fjgj6\") pod \"nova-operator-controller-manager-cfbb9c588-qzxtx\" (UID: \"4cca949a-2334-4d41-bb9e-13e6bd89d7a8\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.309488 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.333233 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-hnczv"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.334317 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.357248 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-xpmwc" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.361266 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.370467 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.378331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kln7p\" (UniqueName: \"kubernetes.io/projected/926db806-d562-44f8-9ac9-71f337dbb202-kube-api-access-kln7p\") pod \"ovn-operator-controller-manager-54fc5f65b7-2jjxl\" (UID: \"926db806-d562-44f8-9ac9-71f337dbb202\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.378374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99379416-bd58-4014-a297-7b14b8ff4aeb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp\" (UID: \"99379416-bd58-4014-a297-7b14b8ff4aeb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.378397 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whv5n\" (UniqueName: \"kubernetes.io/projected/39f22836-fab7-4ffa-b95f-d2dad15a98b9-kube-api-access-whv5n\") pod \"swift-operator-controller-manager-d656998f4-hnczv\" (UID: \"39f22836-fab7-4ffa-b95f-d2dad15a98b9\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.378449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec73cd81-275a-44e0-93dc-6b20123a88d3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-rn29q\" (UID: \"ec73cd81-275a-44e0-93dc-6b20123a88d3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.378465 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tstbb\" (UniqueName: \"kubernetes.io/projected/4676e827-15a7-4286-a4a8-0147b70e315b-kube-api-access-tstbb\") pod \"placement-operator-controller-manager-5b797b8dff-rzg64\" (UID: \"4676e827-15a7-4286-a4a8-0147b70e315b\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.378495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9vk9\" (UniqueName: \"kubernetes.io/projected/3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2-kube-api-access-n9vk9\") pod \"telemetry-operator-controller-manager-6d4bf84b58-w82fg\" (UID: \"3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.378516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz7cd\" (UniqueName: \"kubernetes.io/projected/99379416-bd58-4014-a297-7b14b8ff4aeb-kube-api-access-tz7cd\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp\" (UID: \"99379416-bd58-4014-a297-7b14b8ff4aeb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:24 crc kubenswrapper[4824]: E1124 13:30:24.379353 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 13:30:24 crc kubenswrapper[4824]: E1124 13:30:24.379399 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/99379416-bd58-4014-a297-7b14b8ff4aeb-cert podName:99379416-bd58-4014-a297-7b14b8ff4aeb nodeName:}" failed. No retries permitted until 2025-11-24 13:30:24.879382902 +0000 UTC m=+966.518922212 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/99379416-bd58-4014-a297-7b14b8ff4aeb-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" (UID: "99379416-bd58-4014-a297-7b14b8ff4aeb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.389032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec73cd81-275a-44e0-93dc-6b20123a88d3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-rn29q\" (UID: \"ec73cd81-275a-44e0-93dc-6b20123a88d3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.390282 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-hnczv"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.403101 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.403974 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-fhn99"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.405073 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.415970 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-fhn99"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.423342 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-c6dxt" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.439633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz7cd\" (UniqueName: \"kubernetes.io/projected/99379416-bd58-4014-a297-7b14b8ff4aeb-kube-api-access-tz7cd\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp\" (UID: \"99379416-bd58-4014-a297-7b14b8ff4aeb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.441855 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.441899 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.442702 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kln7p\" (UniqueName: \"kubernetes.io/projected/926db806-d562-44f8-9ac9-71f337dbb202-kube-api-access-kln7p\") pod \"ovn-operator-controller-manager-54fc5f65b7-2jjxl\" (UID: \"926db806-d562-44f8-9ac9-71f337dbb202\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.442918 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.447353 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-jhjwj" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.472870 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.481474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9vk9\" (UniqueName: \"kubernetes.io/projected/3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2-kube-api-access-n9vk9\") pod \"telemetry-operator-controller-manager-6d4bf84b58-w82fg\" (UID: \"3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.481604 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whv5n\" (UniqueName: \"kubernetes.io/projected/39f22836-fab7-4ffa-b95f-d2dad15a98b9-kube-api-access-whv5n\") pod \"swift-operator-controller-manager-d656998f4-hnczv\" (UID: \"39f22836-fab7-4ffa-b95f-d2dad15a98b9\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.481654 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tstbb\" (UniqueName: \"kubernetes.io/projected/4676e827-15a7-4286-a4a8-0147b70e315b-kube-api-access-tstbb\") pod \"placement-operator-controller-manager-5b797b8dff-rzg64\" (UID: \"4676e827-15a7-4286-a4a8-0147b70e315b\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.506992 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.510860 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.554507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tstbb\" (UniqueName: \"kubernetes.io/projected/4676e827-15a7-4286-a4a8-0147b70e315b-kube-api-access-tstbb\") pod \"placement-operator-controller-manager-5b797b8dff-rzg64\" (UID: \"4676e827-15a7-4286-a4a8-0147b70e315b\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.556482 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9vk9\" (UniqueName: \"kubernetes.io/projected/3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2-kube-api-access-n9vk9\") pod \"telemetry-operator-controller-manager-6d4bf84b58-w82fg\" (UID: \"3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.556564 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.567855 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whv5n\" (UniqueName: \"kubernetes.io/projected/39f22836-fab7-4ffa-b95f-d2dad15a98b9-kube-api-access-whv5n\") pod \"swift-operator-controller-manager-d656998f4-hnczv\" (UID: \"39f22836-fab7-4ffa-b95f-d2dad15a98b9\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.585751 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.585858 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.585885 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krqgd\" (UniqueName: \"kubernetes.io/projected/bc237c32-95f2-4386-88f3-a337d2269261-kube-api-access-krqgd\") pod \"test-operator-controller-manager-b4c496f69-fhn99\" (UID: \"bc237c32-95f2-4386-88f3-a337d2269261\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.585922 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtg48\" (UniqueName: \"kubernetes.io/projected/c8001b29-70e7-4861-9b3d-41e732db7bf9-kube-api-access-mtg48\") pod \"watcher-operator-controller-manager-8c6448b9f-t46st\" (UID: \"c8001b29-70e7-4861-9b3d-41e732db7bf9\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.586424 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.595840 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.596043 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ndcbs" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.598061 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.602541 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.603349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.612234 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-hj9sq" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.644638 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.652123 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.690104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krqgd\" (UniqueName: \"kubernetes.io/projected/bc237c32-95f2-4386-88f3-a337d2269261-kube-api-access-krqgd\") pod \"test-operator-controller-manager-b4c496f69-fhn99\" (UID: \"bc237c32-95f2-4386-88f3-a337d2269261\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.690143 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtg48\" (UniqueName: \"kubernetes.io/projected/c8001b29-70e7-4861-9b3d-41e732db7bf9-kube-api-access-mtg48\") pod \"watcher-operator-controller-manager-8c6448b9f-t46st\" (UID: \"c8001b29-70e7-4861-9b3d-41e732db7bf9\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.710397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.739452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtg48\" (UniqueName: \"kubernetes.io/projected/c8001b29-70e7-4861-9b3d-41e732db7bf9-kube-api-access-mtg48\") pod \"watcher-operator-controller-manager-8c6448b9f-t46st\" (UID: \"c8001b29-70e7-4861-9b3d-41e732db7bf9\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.772602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krqgd\" (UniqueName: \"kubernetes.io/projected/bc237c32-95f2-4386-88f3-a337d2269261-kube-api-access-krqgd\") pod \"test-operator-controller-manager-b4c496f69-fhn99\" (UID: \"bc237c32-95f2-4386-88f3-a337d2269261\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.776985 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.794167 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s7lq\" (UniqueName: \"kubernetes.io/projected/c97af640-fa25-46a6-816a-104506790d6d-kube-api-access-6s7lq\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-b82s2\" (UID: \"c97af640-fa25-46a6-816a-104506790d6d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.794210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f304aa2-86ae-45a3-9039-848548cbd500-cert\") pod \"openstack-operator-controller-manager-7547bd9b9-tpbcz\" (UID: \"4f304aa2-86ae-45a3-9039-848548cbd500\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.794288 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pnmm\" (UniqueName: \"kubernetes.io/projected/4f304aa2-86ae-45a3-9039-848548cbd500-kube-api-access-8pnmm\") pod \"openstack-operator-controller-manager-7547bd9b9-tpbcz\" (UID: \"4f304aa2-86ae-45a3-9039-848548cbd500\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.803253 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.909491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s7lq\" (UniqueName: \"kubernetes.io/projected/c97af640-fa25-46a6-816a-104506790d6d-kube-api-access-6s7lq\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-b82s2\" (UID: \"c97af640-fa25-46a6-816a-104506790d6d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.909533 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f304aa2-86ae-45a3-9039-848548cbd500-cert\") pod \"openstack-operator-controller-manager-7547bd9b9-tpbcz\" (UID: \"4f304aa2-86ae-45a3-9039-848548cbd500\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.909591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99379416-bd58-4014-a297-7b14b8ff4aeb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp\" (UID: \"99379416-bd58-4014-a297-7b14b8ff4aeb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.909614 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pnmm\" (UniqueName: \"kubernetes.io/projected/4f304aa2-86ae-45a3-9039-848548cbd500-kube-api-access-8pnmm\") pod \"openstack-operator-controller-manager-7547bd9b9-tpbcz\" (UID: \"4f304aa2-86ae-45a3-9039-848548cbd500\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:24 crc kubenswrapper[4824]: E1124 13:30:24.909849 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 13:30:24 crc kubenswrapper[4824]: E1124 13:30:24.909927 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f304aa2-86ae-45a3-9039-848548cbd500-cert podName:4f304aa2-86ae-45a3-9039-848548cbd500 nodeName:}" failed. No retries permitted until 2025-11-24 13:30:25.409900348 +0000 UTC m=+967.049439658 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4f304aa2-86ae-45a3-9039-848548cbd500-cert") pod "openstack-operator-controller-manager-7547bd9b9-tpbcz" (UID: "4f304aa2-86ae-45a3-9039-848548cbd500") : secret "webhook-server-cert" not found Nov 24 13:30:24 crc kubenswrapper[4824]: E1124 13:30:24.910910 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 13:30:24 crc kubenswrapper[4824]: E1124 13:30:24.910974 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/99379416-bd58-4014-a297-7b14b8ff4aeb-cert podName:99379416-bd58-4014-a297-7b14b8ff4aeb nodeName:}" failed. No retries permitted until 2025-11-24 13:30:25.910954998 +0000 UTC m=+967.550494308 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/99379416-bd58-4014-a297-7b14b8ff4aeb-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" (UID: "99379416-bd58-4014-a297-7b14b8ff4aeb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.940338 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957"] Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.957394 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pnmm\" (UniqueName: \"kubernetes.io/projected/4f304aa2-86ae-45a3-9039-848548cbd500-kube-api-access-8pnmm\") pod \"openstack-operator-controller-manager-7547bd9b9-tpbcz\" (UID: \"4f304aa2-86ae-45a3-9039-848548cbd500\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:24 crc kubenswrapper[4824]: I1124 13:30:24.962036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s7lq\" (UniqueName: \"kubernetes.io/projected/c97af640-fa25-46a6-816a-104506790d6d-kube-api-access-6s7lq\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-b82s2\" (UID: \"c97af640-fa25-46a6-816a-104506790d6d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.055062 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.082889 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.090149 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz"] Nov 24 13:30:25 crc kubenswrapper[4824]: W1124 13:30:25.157683 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f502855_0eb1_455d_8e3b_4af6ac353aa9.slice/crio-1468b0303aef5f67ac1e8c67d7dcf42126f5d3911b002f41fb5cb412842c9f43 WatchSource:0}: Error finding container 1468b0303aef5f67ac1e8c67d7dcf42126f5d3911b002f41fb5cb412842c9f43: Status 404 returned error can't find the container with id 1468b0303aef5f67ac1e8c67d7dcf42126f5d3911b002f41fb5cb412842c9f43 Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.308145 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8"] Nov 24 13:30:25 crc kubenswrapper[4824]: W1124 13:30:25.377980 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88cfde0a_21e8_45ca_a7b3_43699c2ff345.slice/crio-ff5b1147e76514d78886191eacaee11a31357c2de304437f4f98f7ae629c4ff8 WatchSource:0}: Error finding container ff5b1147e76514d78886191eacaee11a31357c2de304437f4f98f7ae629c4ff8: Status 404 returned error can't find the container with id ff5b1147e76514d78886191eacaee11a31357c2de304437f4f98f7ae629c4ff8 Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.418889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f304aa2-86ae-45a3-9039-848548cbd500-cert\") pod \"openstack-operator-controller-manager-7547bd9b9-tpbcz\" (UID: \"4f304aa2-86ae-45a3-9039-848548cbd500\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.425481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f304aa2-86ae-45a3-9039-848548cbd500-cert\") pod \"openstack-operator-controller-manager-7547bd9b9-tpbcz\" (UID: \"4f304aa2-86ae-45a3-9039-848548cbd500\") " pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.460372 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r"] Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.478938 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs"] Nov 24 13:30:25 crc kubenswrapper[4824]: W1124 13:30:25.506018 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9971e72a_bb18_4951_a752_ee96f1725801.slice/crio-596fd95194e3ff47d59fc0db03bc054181683fb365b8f1de5209d36812003a50 WatchSource:0}: Error finding container 596fd95194e3ff47d59fc0db03bc054181683fb365b8f1de5209d36812003a50: Status 404 returned error can't find the container with id 596fd95194e3ff47d59fc0db03bc054181683fb365b8f1de5209d36812003a50 Nov 24 13:30:25 crc kubenswrapper[4824]: W1124 13:30:25.506313 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05a9eb12_a2a1_4fbe_901b_cb817e4713b4.slice/crio-ece4c9e014c144c7cc0d9dbadb6ff7798afeea81f97aeddba0c7cab20d18bef8 WatchSource:0}: Error finding container ece4c9e014c144c7cc0d9dbadb6ff7798afeea81f97aeddba0c7cab20d18bef8: Status 404 returned error can't find the container with id ece4c9e014c144c7cc0d9dbadb6ff7798afeea81f97aeddba0c7cab20d18bef8 Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.513719 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd"] Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.524994 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m"] Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.530398 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.531512 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj"] Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.704571 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q"] Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.715679 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4"] Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.755497 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd"] Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.762245 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6"] Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.769971 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-hnczv"] Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.796142 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" event={"ID":"88cfde0a-21e8-45ca-a7b3-43699c2ff345","Type":"ContainerStarted","Data":"ff5b1147e76514d78886191eacaee11a31357c2de304437f4f98f7ae629c4ff8"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.801144 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" event={"ID":"dc25960c-da62-4fe3-97c1-cf8526bc9c1d","Type":"ContainerStarted","Data":"a56274d53c8cf66754e28ae9f1f3dfc7add3c6c0a66e40cc0df3a6652785a6df"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.802613 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" event={"ID":"50437b59-29ab-4e49-a751-80a83d9e2aa0","Type":"ContainerStarted","Data":"7ae0e97cdcc982fbe480a373fd4ff91ca2117cde148913ab9031f07f2df38d7f"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.803597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" event={"ID":"5f502855-0eb1-455d-8e3b-4af6ac353aa9","Type":"ContainerStarted","Data":"1468b0303aef5f67ac1e8c67d7dcf42126f5d3911b002f41fb5cb412842c9f43"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.804405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" event={"ID":"9971e72a-bb18-4951-a752-ee96f1725801","Type":"ContainerStarted","Data":"596fd95194e3ff47d59fc0db03bc054181683fb365b8f1de5209d36812003a50"} Nov 24 13:30:25 crc kubenswrapper[4824]: W1124 13:30:25.804501 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39f22836_fab7_4ffa_b95f_d2dad15a98b9.slice/crio-051c4e6fcdf8ae7963673aef9e5ff2bdef4e3b852a438af9808a0a1c8f0f0065 WatchSource:0}: Error finding container 051c4e6fcdf8ae7963673aef9e5ff2bdef4e3b852a438af9808a0a1c8f0f0065: Status 404 returned error can't find the container with id 051c4e6fcdf8ae7963673aef9e5ff2bdef4e3b852a438af9808a0a1c8f0f0065 Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.805223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" event={"ID":"9c973000-24b9-4087-a3eb-c5f087a50ae3","Type":"ContainerStarted","Data":"18c765c534a22e5d9c7589a361e5fe8d5c00fe0bbe48afbbef48e48f32f4624e"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.806529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" event={"ID":"3b95ed18-ec44-4712-9f01-adeef55deb75","Type":"ContainerStarted","Data":"5cb19fb53bc2965ec555be7ecb7262f19986d4ac0e9e9701645774210174a7cc"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.816514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" event={"ID":"0cac78b6-e909-4616-bce8-d3ce8ac90452","Type":"ContainerStarted","Data":"d7f81ad27fdb6db4aa78b20b0cd5d30316ee4786bb823bdf19ef0c4e63bb208a"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.824968 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" event={"ID":"3b3da19f-12b1-413d-b6aa-38bc498172c1","Type":"ContainerStarted","Data":"a1a60423ec3e4ee998f45d96620695d8f08a352b5877bc7145148a196abbcfe0"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.830744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" event={"ID":"218ca72b-e49f-4830-8be8-9b6e85956009","Type":"ContainerStarted","Data":"ced4138fe4eae5d0a4d3e81a0eec13c476e2771e1f8bfbeed17108fd0cd13b2b"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.835185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" event={"ID":"65d269bf-f6af-4239-8e51-bf44fe7830d1","Type":"ContainerStarted","Data":"18f2858efaee4c9c2bafef55ef9f7c3252e8859663347a533bb5c8e2da32446c"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.837689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" event={"ID":"ec73cd81-275a-44e0-93dc-6b20123a88d3","Type":"ContainerStarted","Data":"37f8bdd4afe643de959a360b8948cfe135d4f63bafb674af16bbce55ecb0ecbd"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.838788 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" event={"ID":"05a9eb12-a2a1-4fbe-901b-cb817e4713b4","Type":"ContainerStarted","Data":"ece4c9e014c144c7cc0d9dbadb6ff7798afeea81f97aeddba0c7cab20d18bef8"} Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.929379 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99379416-bd58-4014-a297-7b14b8ff4aeb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp\" (UID: \"99379416-bd58-4014-a297-7b14b8ff4aeb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.934185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99379416-bd58-4014-a297-7b14b8ff4aeb-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp\" (UID: \"99379416-bd58-4014-a297-7b14b8ff4aeb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:25 crc kubenswrapper[4824]: I1124 13:30:25.976124 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.090508 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl"] Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.099904 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg"] Nov 24 13:30:26 crc kubenswrapper[4824]: W1124 13:30:26.102760 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc237c32_95f2_4386_88f3_a337d2269261.slice/crio-1f22ddab8add1afdd2a21d0f005b972905afa60fe11a6bead6ed8e7f5ea185ed WatchSource:0}: Error finding container 1f22ddab8add1afdd2a21d0f005b972905afa60fe11a6bead6ed8e7f5ea185ed: Status 404 returned error can't find the container with id 1f22ddab8add1afdd2a21d0f005b972905afa60fe11a6bead6ed8e7f5ea185ed Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.103930 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-fhn99"] Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.116605 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64"] Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.125072 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st"] Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.127516 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tstbb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b797b8dff-rzg64_openstack-operators(4676e827-15a7-4286-a4a8-0147b70e315b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.127651 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-krqgd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-fhn99_openstack-operators(bc237c32-95f2-4386-88f3-a337d2269261): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.148937 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx"] Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.149746 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n9vk9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6d4bf84b58-w82fg_openstack-operators(3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.154751 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2"] Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.155602 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fjgj6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-cfbb9c588-qzxtx_openstack-operators(4cca949a-2334-4d41-bb9e-13e6bd89d7a8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4824]: W1124 13:30:26.160593 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8001b29_70e7_4861_9b3d_41e732db7bf9.slice/crio-8af156d7a5bcfae6548a6df35480c2012357b7171cbe15833c0f056ad4be7f93 WatchSource:0}: Error finding container 8af156d7a5bcfae6548a6df35480c2012357b7171cbe15833c0f056ad4be7f93: Status 404 returned error can't find the container with id 8af156d7a5bcfae6548a6df35480c2012357b7171cbe15833c0f056ad4be7f93 Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.173686 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz"] Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.176108 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mtg48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-t46st_openstack-operators(c8001b29-70e7-4861-9b3d-41e732db7bf9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.479588 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp"] Nov 24 13:30:26 crc kubenswrapper[4824]: W1124 13:30:26.496307 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99379416_bd58_4014_a297_7b14b8ff4aeb.slice/crio-922304fb17e6b362b02c5229324aec0acab647a7e7de2c8ef19430d3f11bf76c WatchSource:0}: Error finding container 922304fb17e6b362b02c5229324aec0acab647a7e7de2c8ef19430d3f11bf76c: Status 404 returned error can't find the container with id 922304fb17e6b362b02c5229324aec0acab647a7e7de2c8ef19430d3f11bf76c Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.572929 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" podUID="4676e827-15a7-4286-a4a8-0147b70e315b" Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.582040 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" podUID="bc237c32-95f2-4386-88f3-a337d2269261" Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.619822 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" podUID="4cca949a-2334-4d41-bb9e-13e6bd89d7a8" Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.654388 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" podUID="c8001b29-70e7-4861-9b3d-41e732db7bf9" Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.663639 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" podUID="3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.858053 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" event={"ID":"c8001b29-70e7-4861-9b3d-41e732db7bf9","Type":"ContainerStarted","Data":"93ef99dadac16efe8e229b77f67a82363e8da5de2e55cc08b93d7f42a8e3e538"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.858103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" event={"ID":"c8001b29-70e7-4861-9b3d-41e732db7bf9","Type":"ContainerStarted","Data":"8af156d7a5bcfae6548a6df35480c2012357b7171cbe15833c0f056ad4be7f93"} Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.860411 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" podUID="c8001b29-70e7-4861-9b3d-41e732db7bf9" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.865190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" event={"ID":"bc237c32-95f2-4386-88f3-a337d2269261","Type":"ContainerStarted","Data":"35901b5fd0a8b9cd5f30b4d1028a08b4cbd9f66aba682f8e1342611129ac0256"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.865232 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" event={"ID":"bc237c32-95f2-4386-88f3-a337d2269261","Type":"ContainerStarted","Data":"1f22ddab8add1afdd2a21d0f005b972905afa60fe11a6bead6ed8e7f5ea185ed"} Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.869430 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" podUID="bc237c32-95f2-4386-88f3-a337d2269261" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.870731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" event={"ID":"c97af640-fa25-46a6-816a-104506790d6d","Type":"ContainerStarted","Data":"157a79a60306c7f40b37fc8c7daebac1345b83701ba0ceb3f84129942b4d263e"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.876630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" event={"ID":"4676e827-15a7-4286-a4a8-0147b70e315b","Type":"ContainerStarted","Data":"bc2612cb42a009ddc4d5269f99bcbe11f5e249859fc2e57d4d33c261792e3fd3"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.876676 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" event={"ID":"4676e827-15a7-4286-a4a8-0147b70e315b","Type":"ContainerStarted","Data":"39183e04a84c27c2c4ed9c897fca203f2fd37468c4a2f0be98d956c572cd9248"} Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.888115 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" podUID="4676e827-15a7-4286-a4a8-0147b70e315b" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.889490 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" event={"ID":"926db806-d562-44f8-9ac9-71f337dbb202","Type":"ContainerStarted","Data":"65717f067d5b3407a3e0750d7b7de51a588a05ae52f658a89567c762319a4d7b"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.917643 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" event={"ID":"4f304aa2-86ae-45a3-9039-848548cbd500","Type":"ContainerStarted","Data":"8527748bf0a7432cfb5b37d0ccd90ba7745ff377abcd79a7127089ed3e6dae18"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.917688 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" event={"ID":"4f304aa2-86ae-45a3-9039-848548cbd500","Type":"ContainerStarted","Data":"0f432f84ca542c82155b5e2160a59e3ae22c68f5e1fda617b78eb1fbed5046c8"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.936039 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" event={"ID":"3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2","Type":"ContainerStarted","Data":"836bc6366670bc4384478077306ad19a006094523fe58b20e910d35fe7e91a3d"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.936082 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" event={"ID":"3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2","Type":"ContainerStarted","Data":"a8eb2eeef445b02a59327d7ab205c8b6bebe9cbc365c245b100fb03fc309e307"} Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.937848 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" podUID="3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.939477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" event={"ID":"99379416-bd58-4014-a297-7b14b8ff4aeb","Type":"ContainerStarted","Data":"922304fb17e6b362b02c5229324aec0acab647a7e7de2c8ef19430d3f11bf76c"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.941263 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" event={"ID":"4cca949a-2334-4d41-bb9e-13e6bd89d7a8","Type":"ContainerStarted","Data":"e6cea6e26bd816d775ece9609c5a7820de1df5ee0152db953a532bac911c390b"} Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.941318 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" event={"ID":"4cca949a-2334-4d41-bb9e-13e6bd89d7a8","Type":"ContainerStarted","Data":"00b4260d4c3e543ec8c3edefde68ac8db05a11fe8e7bea7b7ffb808e5fd24773"} Nov 24 13:30:26 crc kubenswrapper[4824]: E1124 13:30:26.944497 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" podUID="4cca949a-2334-4d41-bb9e-13e6bd89d7a8" Nov 24 13:30:26 crc kubenswrapper[4824]: I1124 13:30:26.949998 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" event={"ID":"39f22836-fab7-4ffa-b95f-d2dad15a98b9","Type":"ContainerStarted","Data":"051c4e6fcdf8ae7963673aef9e5ff2bdef4e3b852a438af9808a0a1c8f0f0065"} Nov 24 13:30:27 crc kubenswrapper[4824]: I1124 13:30:27.966971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" event={"ID":"4f304aa2-86ae-45a3-9039-848548cbd500","Type":"ContainerStarted","Data":"10d1fb158716b402d62080858e2f6cc95d8a8c1bc6bd6c7beac7a205bf54f49f"} Nov 24 13:30:27 crc kubenswrapper[4824]: E1124 13:30:27.969600 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" podUID="4cca949a-2334-4d41-bb9e-13e6bd89d7a8" Nov 24 13:30:27 crc kubenswrapper[4824]: E1124 13:30:27.969665 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" podUID="bc237c32-95f2-4386-88f3-a337d2269261" Nov 24 13:30:27 crc kubenswrapper[4824]: E1124 13:30:27.969752 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" podUID="3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2" Nov 24 13:30:27 crc kubenswrapper[4824]: E1124 13:30:27.969814 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" podUID="4676e827-15a7-4286-a4a8-0147b70e315b" Nov 24 13:30:27 crc kubenswrapper[4824]: E1124 13:30:27.971755 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" podUID="c8001b29-70e7-4861-9b3d-41e732db7bf9" Nov 24 13:30:28 crc kubenswrapper[4824]: I1124 13:30:28.039644 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" podStartSLOduration=4.039607178 podStartE2EDuration="4.039607178s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:30:28.036770128 +0000 UTC m=+969.676309438" watchObservedRunningTime="2025-11-24 13:30:28.039607178 +0000 UTC m=+969.679146488" Nov 24 13:30:28 crc kubenswrapper[4824]: I1124 13:30:28.975991 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:35 crc kubenswrapper[4824]: I1124 13:30:35.540998 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7547bd9b9-tpbcz" Nov 24 13:30:40 crc kubenswrapper[4824]: E1124 13:30:40.206004 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:c6405d94e56b40ef669729216ab4b9c441f34bb280902efa2940038c076b560f" Nov 24 13:30:40 crc kubenswrapper[4824]: E1124 13:30:40.206540 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:c6405d94e56b40ef669729216ab4b9c441f34bb280902efa2940038c076b560f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wlcrm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-767ccfd65f-gw957_openstack-operators(3b95ed18-ec44-4712-9f01-adeef55deb75): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:40 crc kubenswrapper[4824]: E1124 13:30:40.745951 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a" Nov 24 13:30:40 crc kubenswrapper[4824]: E1124 13:30:40.746159 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bxnt6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7454b96578-v6dbs_openstack-operators(65d269bf-f6af-4239-8e51-bf44fe7830d1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:41 crc kubenswrapper[4824]: E1124 13:30:41.192404 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6" Nov 24 13:30:41 crc kubenswrapper[4824]: E1124 13:30:41.193012 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qppbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-78bd47f458-zncj6_openstack-operators(50437b59-29ab-4e49-a751-80a83d9e2aa0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:41 crc kubenswrapper[4824]: E1124 13:30:41.766213 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9" Nov 24 13:30:41 crc kubenswrapper[4824]: E1124 13:30:41.766453 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xfm4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6498cbf48f-h5mx8_openstack-operators(88cfde0a-21e8-45ca-a7b3-43699c2ff345): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:44 crc kubenswrapper[4824]: E1124 13:30:44.855007 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377" Nov 24 13:30:44 crc kubenswrapper[4824]: E1124 13:30:44.855177 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wmqkn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-99b499f4-cmmdj_openstack-operators(05a9eb12-a2a1-4fbe-901b-cb817e4713b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:45 crc kubenswrapper[4824]: E1124 13:30:45.267741 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:d38faa9070da05487afdaa9e261ad39274c2ed862daf42efa460a040431f1991" Nov 24 13:30:45 crc kubenswrapper[4824]: E1124 13:30:45.267966 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:d38faa9070da05487afdaa9e261ad39274c2ed862daf42efa460a040431f1991,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7n9q2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-7969689c84-gvhjd_openstack-operators(9c973000-24b9-4087-a3eb-c5f087a50ae3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:46 crc kubenswrapper[4824]: E1124 13:30:46.477458 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b" Nov 24 13:30:46 crc kubenswrapper[4824]: E1124 13:30:46.477878 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kln7p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-54fc5f65b7-2jjxl_openstack-operators(926db806-d562-44f8-9ac9-71f337dbb202): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:49 crc kubenswrapper[4824]: E1124 13:30:49.399709 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9" Nov 24 13:30:49 crc kubenswrapper[4824]: E1124 13:30:49.400772 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-78gsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-598f69df5d-cdnsz_openstack-operators(5f502855-0eb1-455d-8e3b-4af6ac353aa9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:49 crc kubenswrapper[4824]: E1124 13:30:49.825950 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0" Nov 24 13:30:49 crc kubenswrapper[4824]: E1124 13:30:49.826139 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-whv5n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d656998f4-hnczv_openstack-operators(39f22836-fab7-4ffa-b95f-d2dad15a98b9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:51 crc kubenswrapper[4824]: E1124 13:30:51.537968 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a" Nov 24 13:30:51 crc kubenswrapper[4824]: E1124 13:30:51.538631 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wcm5n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-58f887965d-xdrwd_openstack-operators(3b3da19f-12b1-413d-b6aa-38bc498172c1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:52 crc kubenswrapper[4824]: E1124 13:30:52.051267 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96" Nov 24 13:30:52 crc kubenswrapper[4824]: E1124 13:30:52.051502 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wwflw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-56f54d6746-s2k8r_openstack-operators(0cac78b6-e909-4616-bce8-d3ce8ac90452): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:53 crc kubenswrapper[4824]: E1124 13:30:53.510731 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13" Nov 24 13:30:53 crc kubenswrapper[4824]: E1124 13:30:53.510973 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-stgrc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-54cfbf4c7d-bg2m4_openstack-operators(218ca72b-e49f-4830-8be8-9b6e85956009): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:53 crc kubenswrapper[4824]: E1124 13:30:53.939006 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd" Nov 24 13:30:53 crc kubenswrapper[4824]: E1124 13:30:53.939674 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tz7cd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp_openstack-operators(99379416-bd58-4014-a297-7b14b8ff4aeb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:57 crc kubenswrapper[4824]: E1124 13:30:57.857465 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 24 13:30:57 crc kubenswrapper[4824]: E1124 13:30:57.858006 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6s7lq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-b82s2_openstack-operators(c97af640-fa25-46a6-816a-104506790d6d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:57 crc kubenswrapper[4824]: E1124 13:30:57.859210 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" podUID="c97af640-fa25-46a6-816a-104506790d6d" Nov 24 13:30:58 crc kubenswrapper[4824]: E1124 13:30:58.198424 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" podUID="c97af640-fa25-46a6-816a-104506790d6d" Nov 24 13:30:58 crc kubenswrapper[4824]: E1124 13:30:58.429745 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f" Nov 24 13:30:58 crc kubenswrapper[4824]: E1124 13:30:58.429930 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n9vk9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6d4bf84b58-w82fg_openstack-operators(3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:58 crc kubenswrapper[4824]: E1124 13:30:58.431115 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" podUID="3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2" Nov 24 13:30:58 crc kubenswrapper[4824]: E1124 13:30:58.864374 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d" Nov 24 13:30:58 crc kubenswrapper[4824]: E1124 13:30:58.864531 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-krqgd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-fhn99_openstack-operators(bc237c32-95f2-4386-88f3-a337d2269261): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:30:58 crc kubenswrapper[4824]: E1124 13:30:58.865681 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" podUID="bc237c32-95f2-4386-88f3-a337d2269261" Nov 24 13:31:00 crc kubenswrapper[4824]: E1124 13:31:00.993799 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 24 13:31:00 crc kubenswrapper[4824]: E1124 13:31:00.994333 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mtg48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-t46st_openstack-operators(c8001b29-70e7-4861-9b3d-41e732db7bf9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:31:00 crc kubenswrapper[4824]: E1124 13:31:00.995625 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" podUID="c8001b29-70e7-4861-9b3d-41e732db7bf9" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.640650 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" podUID="50437b59-29ab-4e49-a751-80a83d9e2aa0" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.659328 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" podUID="88cfde0a-21e8-45ca-a7b3-43699c2ff345" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.664395 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" podUID="218ca72b-e49f-4830-8be8-9b6e85956009" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.664552 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" podUID="99379416-bd58-4014-a297-7b14b8ff4aeb" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.664698 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" podUID="05a9eb12-a2a1-4fbe-901b-cb817e4713b4" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.667979 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" podUID="9c973000-24b9-4087-a3eb-c5f087a50ae3" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.671107 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" podUID="65d269bf-f6af-4239-8e51-bf44fe7830d1" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.678116 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" podUID="39f22836-fab7-4ffa-b95f-d2dad15a98b9" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.678245 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" podUID="926db806-d562-44f8-9ac9-71f337dbb202" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.681733 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" podUID="3b95ed18-ec44-4712-9f01-adeef55deb75" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.765820 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" podUID="5f502855-0eb1-455d-8e3b-4af6ac353aa9" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.787509 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" podUID="3b3da19f-12b1-413d-b6aa-38bc498172c1" Nov 24 13:31:01 crc kubenswrapper[4824]: E1124 13:31:01.918034 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" podUID="0cac78b6-e909-4616-bce8-d3ce8ac90452" Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.220029 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" event={"ID":"4cca949a-2334-4d41-bb9e-13e6bd89d7a8","Type":"ContainerStarted","Data":"3223f9581beed6416e413805f31b15f9642c7e9bce41572df783dcddbbfa8d3a"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.220636 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.222113 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" event={"ID":"88cfde0a-21e8-45ca-a7b3-43699c2ff345","Type":"ContainerStarted","Data":"90dd04b4db8719177e5f3d08fa110f3112e772ba96c804c2d3d7b8daffefac3d"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.223777 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" event={"ID":"ec73cd81-275a-44e0-93dc-6b20123a88d3","Type":"ContainerStarted","Data":"1c9fa0083cfb148f7df8d81380493eae8730e662c483939dbc4aeee5a35b4f32"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.225110 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" event={"ID":"50437b59-29ab-4e49-a751-80a83d9e2aa0","Type":"ContainerStarted","Data":"06e264da8a5a53959267ac205e92deefcf2d781ea05b016a18951c2a28c600ce"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.228694 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" event={"ID":"dc25960c-da62-4fe3-97c1-cf8526bc9c1d","Type":"ContainerStarted","Data":"c92537483b786b2d6cead47c7e8c4185324b7284e37893f43852fab3b0c5ea1b"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.228824 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" event={"ID":"dc25960c-da62-4fe3-97c1-cf8526bc9c1d","Type":"ContainerStarted","Data":"4207783a749ea3598b48f7a86cc346a420ed5562cb2b3f15c90dc495fb0ba864"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.231255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" event={"ID":"99379416-bd58-4014-a297-7b14b8ff4aeb","Type":"ContainerStarted","Data":"9b0c449e2fca39bdbcea234bff6f055a3c8e91f8de3d94d5a1b7224ff4ee28f9"} Nov 24 13:31:02 crc kubenswrapper[4824]: E1124 13:31:02.233776 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" podUID="99379416-bd58-4014-a297-7b14b8ff4aeb" Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.245947 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" event={"ID":"218ca72b-e49f-4830-8be8-9b6e85956009","Type":"ContainerStarted","Data":"3624df23089333bcff3dd16d86ab8041688539d783129fcc2e4d7982a8239a38"} Nov 24 13:31:02 crc kubenswrapper[4824]: E1124 13:31:02.247338 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" podUID="218ca72b-e49f-4830-8be8-9b6e85956009" Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.253373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" event={"ID":"0cac78b6-e909-4616-bce8-d3ce8ac90452","Type":"ContainerStarted","Data":"7b113828afc6ef7d8cdc7721e18226c53eb83d25d52795cf54da3e38a8a5b442"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.255755 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" event={"ID":"4676e827-15a7-4286-a4a8-0147b70e315b","Type":"ContainerStarted","Data":"0b4febbc34fb612f0fa6e62c652e0c5354f572133136a27a86b85891414acec6"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.256234 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.264302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" event={"ID":"926db806-d562-44f8-9ac9-71f337dbb202","Type":"ContainerStarted","Data":"fe64a526d864fcfd4bcc8feb4177df77460668546564c537e1b11752278aecfd"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.272084 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" event={"ID":"3b3da19f-12b1-413d-b6aa-38bc498172c1","Type":"ContainerStarted","Data":"af48ff41943246cf127388640151af1582fe591c9232997f9585683f472e2861"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.279192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" event={"ID":"9c973000-24b9-4087-a3eb-c5f087a50ae3","Type":"ContainerStarted","Data":"5d4c2f788698abbe5afb67b907b073f5d54213f28ce17d73df93a6066ef4c714"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.285968 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" event={"ID":"3b95ed18-ec44-4712-9f01-adeef55deb75","Type":"ContainerStarted","Data":"b33d15395fde088c50b1610ce829b770dd4dcb1d528f93d5b5a9ae21b149ce14"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.297046 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" podStartSLOduration=5.041045409 podStartE2EDuration="39.297026056s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.155427979 +0000 UTC m=+967.794967289" lastFinishedPulling="2025-11-24 13:31:00.411408616 +0000 UTC m=+1002.050947936" observedRunningTime="2025-11-24 13:31:02.290923117 +0000 UTC m=+1003.930462427" watchObservedRunningTime="2025-11-24 13:31:02.297026056 +0000 UTC m=+1003.936565366" Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.297063 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" event={"ID":"5f502855-0eb1-455d-8e3b-4af6ac353aa9","Type":"ContainerStarted","Data":"839edb90e84f7e1cb7ed821e2845f1c9bcacc05080b0954953ec124a2d52f6ff"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.302597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" event={"ID":"9971e72a-bb18-4951-a752-ee96f1725801","Type":"ContainerStarted","Data":"df3f2b1d55cdedae331301c3f3c82eee27bd676972a2f06efcba6527be0c18e9"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.313134 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" event={"ID":"65d269bf-f6af-4239-8e51-bf44fe7830d1","Type":"ContainerStarted","Data":"c179e3496adeb110e6bb018f78eaa6d9fe49bba3759b42872be43bd585aad184"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.318122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" event={"ID":"39f22836-fab7-4ffa-b95f-d2dad15a98b9","Type":"ContainerStarted","Data":"c4ace3e06d62681196dfc75e24f678ba6d6ee2039b39762bb0da99f683ad09ec"} Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.327194 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" event={"ID":"05a9eb12-a2a1-4fbe-901b-cb817e4713b4","Type":"ContainerStarted","Data":"21f7dc58618aff9ef446168a44b67e1e7bacc63a718013ddb5b4887a7f3a62dc"} Nov 24 13:31:02 crc kubenswrapper[4824]: E1124 13:31:02.331827 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wmqkn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-99b499f4-cmmdj_openstack-operators(05a9eb12-a2a1-4fbe-901b-cb817e4713b4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:31:02 crc kubenswrapper[4824]: E1124 13:31:02.334408 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" podUID="05a9eb12-a2a1-4fbe-901b-cb817e4713b4" Nov 24 13:31:02 crc kubenswrapper[4824]: I1124 13:31:02.658382 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" podStartSLOduration=4.787282788 podStartE2EDuration="39.658360644s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.12734845 +0000 UTC m=+967.766887760" lastFinishedPulling="2025-11-24 13:31:00.998426306 +0000 UTC m=+1002.637965616" observedRunningTime="2025-11-24 13:31:02.655730296 +0000 UTC m=+1004.295269616" watchObservedRunningTime="2025-11-24 13:31:02.658360644 +0000 UTC m=+1004.297899954" Nov 24 13:31:03 crc kubenswrapper[4824]: I1124 13:31:03.335434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" event={"ID":"9971e72a-bb18-4951-a752-ee96f1725801","Type":"ContainerStarted","Data":"8579fa846cf24d065a59a852019d0d79787f5727c2043c4ed7e732306e372bf9"} Nov 24 13:31:03 crc kubenswrapper[4824]: I1124 13:31:03.337346 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" event={"ID":"ec73cd81-275a-44e0-93dc-6b20123a88d3","Type":"ContainerStarted","Data":"c3a2e3d6032c7236ca24fd948267f4dba54e3ba9d5509efa3b092784bc339e31"} Nov 24 13:31:03 crc kubenswrapper[4824]: I1124 13:31:03.338204 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" Nov 24 13:31:03 crc kubenswrapper[4824]: E1124 13:31:03.339112 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" podUID="99379416-bd58-4014-a297-7b14b8ff4aeb" Nov 24 13:31:03 crc kubenswrapper[4824]: E1124 13:31:03.339125 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" podUID="218ca72b-e49f-4830-8be8-9b6e85956009" Nov 24 13:31:03 crc kubenswrapper[4824]: E1124 13:31:03.339176 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" podUID="05a9eb12-a2a1-4fbe-901b-cb817e4713b4" Nov 24 13:31:03 crc kubenswrapper[4824]: I1124 13:31:03.356326 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" podStartSLOduration=6.885711774 podStartE2EDuration="40.356311267s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:24.952084894 +0000 UTC m=+966.591624194" lastFinishedPulling="2025-11-24 13:30:58.422684377 +0000 UTC m=+1000.062223687" observedRunningTime="2025-11-24 13:31:03.351258536 +0000 UTC m=+1004.990797846" watchObservedRunningTime="2025-11-24 13:31:03.356311267 +0000 UTC m=+1004.995850577" Nov 24 13:31:04 crc kubenswrapper[4824]: I1124 13:31:04.344987 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" Nov 24 13:31:04 crc kubenswrapper[4824]: I1124 13:31:04.346405 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:31:04 crc kubenswrapper[4824]: I1124 13:31:04.363676 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" podStartSLOduration=8.448218843 podStartE2EDuration="41.363655304s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.508300854 +0000 UTC m=+967.147840164" lastFinishedPulling="2025-11-24 13:30:58.423737295 +0000 UTC m=+1000.063276625" observedRunningTime="2025-11-24 13:31:04.361382605 +0000 UTC m=+1006.000921915" watchObservedRunningTime="2025-11-24 13:31:04.363655304 +0000 UTC m=+1006.003194614" Nov 24 13:31:04 crc kubenswrapper[4824]: I1124 13:31:04.391551 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" podStartSLOduration=12.745641297 podStartE2EDuration="41.391533501s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.751029409 +0000 UTC m=+967.390568719" lastFinishedPulling="2025-11-24 13:30:54.396921613 +0000 UTC m=+996.036460923" observedRunningTime="2025-11-24 13:31:04.389935099 +0000 UTC m=+1006.029474439" watchObservedRunningTime="2025-11-24 13:31:04.391533501 +0000 UTC m=+1006.031072821" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.351476 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" event={"ID":"3b3da19f-12b1-413d-b6aa-38bc498172c1","Type":"ContainerStarted","Data":"0aad587872fe6ec15ed086bbbba66f438f57f630e1571b754fd6a16a5e695c82"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.351560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.353514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" event={"ID":"5f502855-0eb1-455d-8e3b-4af6ac353aa9","Type":"ContainerStarted","Data":"23c48f073fdc8da5cf07a616c6dfdc0388bbf6c9d727453bb7e37deeb0e64807"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.353617 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.355115 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" event={"ID":"65d269bf-f6af-4239-8e51-bf44fe7830d1","Type":"ContainerStarted","Data":"c651860ae5a3937b180236ddb9a2a414e5475ab7eff0d6ef461acd1e72d0ca62"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.355461 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.357276 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" event={"ID":"3b95ed18-ec44-4712-9f01-adeef55deb75","Type":"ContainerStarted","Data":"ba8bd43f3ba9b8ffc1d82b7d0f5ad543bedd36dfb2e8e2e2a20ead3ca9d9856d"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.357626 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.359932 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" event={"ID":"926db806-d562-44f8-9ac9-71f337dbb202","Type":"ContainerStarted","Data":"4c225901c543d3ff1e548fab22a4bd1ed057297541bed05e25abcb72e2854460"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.360316 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.361765 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" event={"ID":"0cac78b6-e909-4616-bce8-d3ce8ac90452","Type":"ContainerStarted","Data":"66d86b12b22f3a2d16156d485fb9dcc7891aa92cbb9c278ce19c0b4799530f96"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.362196 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.363860 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" event={"ID":"50437b59-29ab-4e49-a751-80a83d9e2aa0","Type":"ContainerStarted","Data":"d3418876c0d8316cb68c3f7abc6715729f9e1089a2b73480382b178c62846f2c"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.364190 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.365237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" event={"ID":"88cfde0a-21e8-45ca-a7b3-43699c2ff345","Type":"ContainerStarted","Data":"fbf318c102e59fc8dc4a795abe54b9a12b3b8a8ab8f31e873c538b92352400cf"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.365579 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.366974 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" event={"ID":"39f22836-fab7-4ffa-b95f-d2dad15a98b9","Type":"ContainerStarted","Data":"c2747046b81d5a629fcadb1e2a5f1b7c14aa106a61970705bb619693c06068ee"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.367341 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.369256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" event={"ID":"9c973000-24b9-4087-a3eb-c5f087a50ae3","Type":"ContainerStarted","Data":"5c5e68cb5a3b7f2f2ec35ad39e723d6e6c42d898a8a133966f217fd14ea60d2f"} Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.369279 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.374460 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" podStartSLOduration=3.600700453 podStartE2EDuration="42.374448792s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.788462211 +0000 UTC m=+967.428001511" lastFinishedPulling="2025-11-24 13:31:04.56221054 +0000 UTC m=+1006.201749850" observedRunningTime="2025-11-24 13:31:05.371323861 +0000 UTC m=+1007.010863171" watchObservedRunningTime="2025-11-24 13:31:05.374448792 +0000 UTC m=+1007.013988102" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.397078 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" podStartSLOduration=3.4726122249999998 podStartE2EDuration="42.397058211s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.511112673 +0000 UTC m=+967.150651983" lastFinishedPulling="2025-11-24 13:31:04.435558659 +0000 UTC m=+1006.075097969" observedRunningTime="2025-11-24 13:31:05.390302285 +0000 UTC m=+1007.029841595" watchObservedRunningTime="2025-11-24 13:31:05.397058211 +0000 UTC m=+1007.036597521" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.420383 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" podStartSLOduration=3.925027932 podStartE2EDuration="42.420363829s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.126953559 +0000 UTC m=+967.766492869" lastFinishedPulling="2025-11-24 13:31:04.622289446 +0000 UTC m=+1006.261828766" observedRunningTime="2025-11-24 13:31:05.414982929 +0000 UTC m=+1007.054522239" watchObservedRunningTime="2025-11-24 13:31:05.420363829 +0000 UTC m=+1007.059903139" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.433565 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" podStartSLOduration=3.50180556 podStartE2EDuration="42.433546623s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.502497841 +0000 UTC m=+967.142037151" lastFinishedPulling="2025-11-24 13:31:04.434238904 +0000 UTC m=+1006.073778214" observedRunningTime="2025-11-24 13:31:05.43304756 +0000 UTC m=+1007.072586870" watchObservedRunningTime="2025-11-24 13:31:05.433546623 +0000 UTC m=+1007.073085933" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.460560 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" podStartSLOduration=4.056649879 podStartE2EDuration="42.460537806s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.788550114 +0000 UTC m=+967.428089424" lastFinishedPulling="2025-11-24 13:31:04.192438041 +0000 UTC m=+1005.831977351" observedRunningTime="2025-11-24 13:31:05.450374781 +0000 UTC m=+1007.089914091" watchObservedRunningTime="2025-11-24 13:31:05.460537806 +0000 UTC m=+1007.100077116" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.470293 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" podStartSLOduration=3.099522447 podStartE2EDuration="42.47027451s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.192470133 +0000 UTC m=+966.832009443" lastFinishedPulling="2025-11-24 13:31:04.563222196 +0000 UTC m=+1006.202761506" observedRunningTime="2025-11-24 13:31:05.467491357 +0000 UTC m=+1007.107030677" watchObservedRunningTime="2025-11-24 13:31:05.47027451 +0000 UTC m=+1007.109813820" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.499492 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" podStartSLOduration=3.131769749 podStartE2EDuration="42.499472811s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.398251899 +0000 UTC m=+967.037791209" lastFinishedPulling="2025-11-24 13:31:04.765954961 +0000 UTC m=+1006.405494271" observedRunningTime="2025-11-24 13:31:05.498367472 +0000 UTC m=+1007.137906782" watchObservedRunningTime="2025-11-24 13:31:05.499472811 +0000 UTC m=+1007.139012121" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.534712 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" podStartSLOduration=3.144649671 podStartE2EDuration="42.534692529s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.045186172 +0000 UTC m=+966.684725482" lastFinishedPulling="2025-11-24 13:31:04.43522902 +0000 UTC m=+1006.074768340" observedRunningTime="2025-11-24 13:31:05.526592568 +0000 UTC m=+1007.166131888" watchObservedRunningTime="2025-11-24 13:31:05.534692529 +0000 UTC m=+1007.174231839" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.574545 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" podStartSLOduration=3.428836429 podStartE2EDuration="42.574527637s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.484236727 +0000 UTC m=+967.123776037" lastFinishedPulling="2025-11-24 13:31:04.629927935 +0000 UTC m=+1006.269467245" observedRunningTime="2025-11-24 13:31:05.556078967 +0000 UTC m=+1007.195618277" watchObservedRunningTime="2025-11-24 13:31:05.574527637 +0000 UTC m=+1007.214066937" Nov 24 13:31:05 crc kubenswrapper[4824]: I1124 13:31:05.575703 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" podStartSLOduration=3.683614952 podStartE2EDuration="42.575696868s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.806651653 +0000 UTC m=+967.446190963" lastFinishedPulling="2025-11-24 13:31:04.698733549 +0000 UTC m=+1006.338272879" observedRunningTime="2025-11-24 13:31:05.574486486 +0000 UTC m=+1007.214025806" watchObservedRunningTime="2025-11-24 13:31:05.575696868 +0000 UTC m=+1007.215236178" Nov 24 13:31:10 crc kubenswrapper[4824]: I1124 13:31:10.787582 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:31:10 crc kubenswrapper[4824]: I1124 13:31:10.787927 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:31:12 crc kubenswrapper[4824]: E1124 13:31:12.015730 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" podUID="c8001b29-70e7-4861-9b3d-41e732db7bf9" Nov 24 13:31:12 crc kubenswrapper[4824]: E1124 13:31:12.015758 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" podUID="bc237c32-95f2-4386-88f3-a337d2269261" Nov 24 13:31:13 crc kubenswrapper[4824]: E1124 13:31:13.012013 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" podUID="3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2" Nov 24 13:31:13 crc kubenswrapper[4824]: I1124 13:31:13.432088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" event={"ID":"c97af640-fa25-46a6-816a-104506790d6d","Type":"ContainerStarted","Data":"625033e9ad6061ccf24c97bca69fdacc3051a7ba4d7a81abfaeae32bd01705db"} Nov 24 13:31:13 crc kubenswrapper[4824]: I1124 13:31:13.452490 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b82s2" podStartSLOduration=3.087800892 podStartE2EDuration="49.452463243s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.155136121 +0000 UTC m=+967.794675431" lastFinishedPulling="2025-11-24 13:31:12.519798472 +0000 UTC m=+1014.159337782" observedRunningTime="2025-11-24 13:31:13.449197878 +0000 UTC m=+1015.088737288" watchObservedRunningTime="2025-11-24 13:31:13.452463243 +0000 UTC m=+1015.092002603" Nov 24 13:31:13 crc kubenswrapper[4824]: I1124 13:31:13.703714 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-747cg" Nov 24 13:31:13 crc kubenswrapper[4824]: I1124 13:31:13.716145 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-h5mx8" Nov 24 13:31:13 crc kubenswrapper[4824]: I1124 13:31:13.781371 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-gw957" Nov 24 13:31:13 crc kubenswrapper[4824]: I1124 13:31:13.803638 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-gvhjd" Nov 24 13:31:13 crc kubenswrapper[4824]: I1124 13:31:13.854634 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-s2k8r" Nov 24 13:31:13 crc kubenswrapper[4824]: I1124 13:31:13.879870 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-cdnsz" Nov 24 13:31:14 crc kubenswrapper[4824]: E1124 13:31:14.011774 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" podUID="05a9eb12-a2a1-4fbe-901b-cb817e4713b4" Nov 24 13:31:14 crc kubenswrapper[4824]: I1124 13:31:14.096444 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-v6dbs" Nov 24 13:31:14 crc kubenswrapper[4824]: I1124 13:31:14.139517 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-mzs9m" Nov 24 13:31:14 crc kubenswrapper[4824]: I1124 13:31:14.406923 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-xdrwd" Nov 24 13:31:14 crc kubenswrapper[4824]: I1124 13:31:14.446088 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-zncj6" Nov 24 13:31:14 crc kubenswrapper[4824]: I1124 13:31:14.510697 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-2jjxl" Nov 24 13:31:14 crc kubenswrapper[4824]: I1124 13:31:14.517619 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-rn29q" Nov 24 13:31:14 crc kubenswrapper[4824]: I1124 13:31:14.590359 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-qzxtx" Nov 24 13:31:14 crc kubenswrapper[4824]: I1124 13:31:14.606688 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-rzg64" Nov 24 13:31:14 crc kubenswrapper[4824]: I1124 13:31:14.713453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-hnczv" Nov 24 13:31:16 crc kubenswrapper[4824]: I1124 13:31:16.471475 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" event={"ID":"99379416-bd58-4014-a297-7b14b8ff4aeb","Type":"ContainerStarted","Data":"9203ad0413ff3796ef13bddf72e162544b2ec47b1c3a6993b2ad72bb0c6d3d07"} Nov 24 13:31:16 crc kubenswrapper[4824]: I1124 13:31:16.515001 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" podStartSLOduration=4.513297538 podStartE2EDuration="53.514982841s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.498591218 +0000 UTC m=+968.138130528" lastFinishedPulling="2025-11-24 13:31:15.500276521 +0000 UTC m=+1017.139815831" observedRunningTime="2025-11-24 13:31:16.507534216 +0000 UTC m=+1018.147073536" watchObservedRunningTime="2025-11-24 13:31:16.514982841 +0000 UTC m=+1018.154522171" Nov 24 13:31:18 crc kubenswrapper[4824]: I1124 13:31:18.492714 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" event={"ID":"218ca72b-e49f-4830-8be8-9b6e85956009","Type":"ContainerStarted","Data":"74e5e8a49671811a601886474aa325e0cb9c4933a37eb7a0cedfb2da3e86fdeb"} Nov 24 13:31:18 crc kubenswrapper[4824]: I1124 13:31:18.493419 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" Nov 24 13:31:18 crc kubenswrapper[4824]: I1124 13:31:18.517398 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" podStartSLOduration=3.794026815 podStartE2EDuration="55.517376664s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.741552892 +0000 UTC m=+967.381092202" lastFinishedPulling="2025-11-24 13:31:17.464902741 +0000 UTC m=+1019.104442051" observedRunningTime="2025-11-24 13:31:18.514249753 +0000 UTC m=+1020.153789143" watchObservedRunningTime="2025-11-24 13:31:18.517376664 +0000 UTC m=+1020.156915984" Nov 24 13:31:24 crc kubenswrapper[4824]: I1124 13:31:24.263868 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-bg2m4" Nov 24 13:31:25 crc kubenswrapper[4824]: I1124 13:31:25.012315 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:31:25 crc kubenswrapper[4824]: I1124 13:31:25.976772 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:31:25 crc kubenswrapper[4824]: I1124 13:31:25.982363 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp" Nov 24 13:31:26 crc kubenswrapper[4824]: I1124 13:31:26.547677 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" event={"ID":"c8001b29-70e7-4861-9b3d-41e732db7bf9","Type":"ContainerStarted","Data":"ae060c1ee290413134f320fe694a8bb02f67fbee5edee5d8c5888c486ac56123"} Nov 24 13:31:26 crc kubenswrapper[4824]: I1124 13:31:26.548358 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" Nov 24 13:31:26 crc kubenswrapper[4824]: I1124 13:31:26.550949 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" event={"ID":"3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2","Type":"ContainerStarted","Data":"94b4da22d23ecb381d53b94e3ae063b9bb9eba384ec0ab7b41231337011c6197"} Nov 24 13:31:26 crc kubenswrapper[4824]: I1124 13:31:26.551749 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" Nov 24 13:31:26 crc kubenswrapper[4824]: I1124 13:31:26.591569 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" podStartSLOduration=3.293087283 podStartE2EDuration="1m2.591542655s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.175955166 +0000 UTC m=+967.815494466" lastFinishedPulling="2025-11-24 13:31:25.474410518 +0000 UTC m=+1027.113949838" observedRunningTime="2025-11-24 13:31:26.575293542 +0000 UTC m=+1028.214832852" watchObservedRunningTime="2025-11-24 13:31:26.591542655 +0000 UTC m=+1028.231081985" Nov 24 13:31:26 crc kubenswrapper[4824]: I1124 13:31:26.596862 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" podStartSLOduration=4.26972313 podStartE2EDuration="1m3.596841813s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.149335358 +0000 UTC m=+967.788874668" lastFinishedPulling="2025-11-24 13:31:25.476454041 +0000 UTC m=+1027.115993351" observedRunningTime="2025-11-24 13:31:26.590355895 +0000 UTC m=+1028.229895215" watchObservedRunningTime="2025-11-24 13:31:26.596841813 +0000 UTC m=+1028.236381153" Nov 24 13:31:27 crc kubenswrapper[4824]: I1124 13:31:27.557871 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" event={"ID":"bc237c32-95f2-4386-88f3-a337d2269261","Type":"ContainerStarted","Data":"07235f24ec49047421b2c1ac10b4e458ae7b073c20051bd5ee11dd977cac7d11"} Nov 24 13:31:27 crc kubenswrapper[4824]: I1124 13:31:27.559066 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" Nov 24 13:31:27 crc kubenswrapper[4824]: I1124 13:31:27.561207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" event={"ID":"05a9eb12-a2a1-4fbe-901b-cb817e4713b4","Type":"ContainerStarted","Data":"f5af69a4cf6949cb17b11a4c5c98a8805dd4295462da4cb4df88d3ef17694353"} Nov 24 13:31:27 crc kubenswrapper[4824]: I1124 13:31:27.561676 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" Nov 24 13:31:27 crc kubenswrapper[4824]: I1124 13:31:27.607453 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" podStartSLOduration=2.959773598 podStartE2EDuration="1m3.607433416s" podCreationTimestamp="2025-11-24 13:30:24 +0000 UTC" firstStartedPulling="2025-11-24 13:30:26.127585276 +0000 UTC m=+967.767124586" lastFinishedPulling="2025-11-24 13:31:26.775245054 +0000 UTC m=+1028.414784404" observedRunningTime="2025-11-24 13:31:27.603228366 +0000 UTC m=+1029.242767696" watchObservedRunningTime="2025-11-24 13:31:27.607433416 +0000 UTC m=+1029.246972746" Nov 24 13:31:27 crc kubenswrapper[4824]: I1124 13:31:27.627954 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" podStartSLOduration=3.615271826 podStartE2EDuration="1m4.6279392s" podCreationTimestamp="2025-11-24 13:30:23 +0000 UTC" firstStartedPulling="2025-11-24 13:30:25.511782402 +0000 UTC m=+967.151321712" lastFinishedPulling="2025-11-24 13:31:26.524449776 +0000 UTC m=+1028.163989086" observedRunningTime="2025-11-24 13:31:27.621217375 +0000 UTC m=+1029.260756695" watchObservedRunningTime="2025-11-24 13:31:27.6279392 +0000 UTC m=+1029.267478510" Nov 24 13:31:34 crc kubenswrapper[4824]: I1124 13:31:34.280246 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-cmmdj" Nov 24 13:31:34 crc kubenswrapper[4824]: I1124 13:31:34.657381 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-w82fg" Nov 24 13:31:34 crc kubenswrapper[4824]: I1124 13:31:34.809553 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-t46st" Nov 24 13:31:35 crc kubenswrapper[4824]: I1124 13:31:35.058978 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-fhn99" Nov 24 13:31:40 crc kubenswrapper[4824]: I1124 13:31:40.789523 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:31:40 crc kubenswrapper[4824]: I1124 13:31:40.790233 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:31:53 crc kubenswrapper[4824]: I1124 13:31:53.987742 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-895h4"] Nov 24 13:31:53 crc kubenswrapper[4824]: I1124 13:31:53.989414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:31:53 crc kubenswrapper[4824]: I1124 13:31:53.991773 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 13:31:53 crc kubenswrapper[4824]: I1124 13:31:53.991959 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 13:31:53 crc kubenswrapper[4824]: I1124 13:31:53.992141 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-dhg8l" Nov 24 13:31:53 crc kubenswrapper[4824]: I1124 13:31:53.994600 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.003580 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-895h4"] Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.032502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p769b\" (UniqueName: \"kubernetes.io/projected/9f81886c-379a-4e1d-a5bb-961cbf65927c-kube-api-access-p769b\") pod \"dnsmasq-dns-675f4bcbfc-895h4\" (UID: \"9f81886c-379a-4e1d-a5bb-961cbf65927c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.032549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f81886c-379a-4e1d-a5bb-961cbf65927c-config\") pod \"dnsmasq-dns-675f4bcbfc-895h4\" (UID: \"9f81886c-379a-4e1d-a5bb-961cbf65927c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.078397 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vmsqq"] Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.079625 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.087639 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.090923 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vmsqq"] Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.134874 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p769b\" (UniqueName: \"kubernetes.io/projected/9f81886c-379a-4e1d-a5bb-961cbf65927c-kube-api-access-p769b\") pod \"dnsmasq-dns-675f4bcbfc-895h4\" (UID: \"9f81886c-379a-4e1d-a5bb-961cbf65927c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.134920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-config\") pod \"dnsmasq-dns-78dd6ddcc-vmsqq\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.134936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd8ww\" (UniqueName: \"kubernetes.io/projected/041daaa0-0f4f-45aa-bda7-589182e65764-kube-api-access-zd8ww\") pod \"dnsmasq-dns-78dd6ddcc-vmsqq\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.134961 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f81886c-379a-4e1d-a5bb-961cbf65927c-config\") pod \"dnsmasq-dns-675f4bcbfc-895h4\" (UID: \"9f81886c-379a-4e1d-a5bb-961cbf65927c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.134999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vmsqq\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.135921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f81886c-379a-4e1d-a5bb-961cbf65927c-config\") pod \"dnsmasq-dns-675f4bcbfc-895h4\" (UID: \"9f81886c-379a-4e1d-a5bb-961cbf65927c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.160856 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p769b\" (UniqueName: \"kubernetes.io/projected/9f81886c-379a-4e1d-a5bb-961cbf65927c-kube-api-access-p769b\") pod \"dnsmasq-dns-675f4bcbfc-895h4\" (UID: \"9f81886c-379a-4e1d-a5bb-961cbf65927c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.236079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vmsqq\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.236172 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-config\") pod \"dnsmasq-dns-78dd6ddcc-vmsqq\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.236192 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd8ww\" (UniqueName: \"kubernetes.io/projected/041daaa0-0f4f-45aa-bda7-589182e65764-kube-api-access-zd8ww\") pod \"dnsmasq-dns-78dd6ddcc-vmsqq\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.237205 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vmsqq\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.237316 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-config\") pod \"dnsmasq-dns-78dd6ddcc-vmsqq\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.264826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd8ww\" (UniqueName: \"kubernetes.io/projected/041daaa0-0f4f-45aa-bda7-589182e65764-kube-api-access-zd8ww\") pod \"dnsmasq-dns-78dd6ddcc-vmsqq\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.312578 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.400482 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.792864 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-895h4"] Nov 24 13:31:54 crc kubenswrapper[4824]: I1124 13:31:54.867438 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vmsqq"] Nov 24 13:31:54 crc kubenswrapper[4824]: W1124 13:31:54.872980 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod041daaa0_0f4f_45aa_bda7_589182e65764.slice/crio-aee4dd1676f49ad1367bf4664d47c4472f4d91674ad1d3c032ddda43ef2742bb WatchSource:0}: Error finding container aee4dd1676f49ad1367bf4664d47c4472f4d91674ad1d3c032ddda43ef2742bb: Status 404 returned error can't find the container with id aee4dd1676f49ad1367bf4664d47c4472f4d91674ad1d3c032ddda43ef2742bb Nov 24 13:31:55 crc kubenswrapper[4824]: I1124 13:31:55.776487 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" event={"ID":"041daaa0-0f4f-45aa-bda7-589182e65764","Type":"ContainerStarted","Data":"aee4dd1676f49ad1367bf4664d47c4472f4d91674ad1d3c032ddda43ef2742bb"} Nov 24 13:31:55 crc kubenswrapper[4824]: I1124 13:31:55.777605 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" event={"ID":"9f81886c-379a-4e1d-a5bb-961cbf65927c","Type":"ContainerStarted","Data":"42d413c2f9956dce31ad601e22f03c57fbd00f76628ed0eea6856604ba86291f"} Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.591134 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-895h4"] Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.612487 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sxgbr"] Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.613736 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.637157 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sxgbr"] Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.688713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd4hb\" (UniqueName: \"kubernetes.io/projected/7456224a-635c-4191-ad12-e7ecd14192e6-kube-api-access-cd4hb\") pod \"dnsmasq-dns-666b6646f7-sxgbr\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.688783 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-config\") pod \"dnsmasq-dns-666b6646f7-sxgbr\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.688800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-sxgbr\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.792401 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd4hb\" (UniqueName: \"kubernetes.io/projected/7456224a-635c-4191-ad12-e7ecd14192e6-kube-api-access-cd4hb\") pod \"dnsmasq-dns-666b6646f7-sxgbr\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.792466 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-sxgbr\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.792490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-config\") pod \"dnsmasq-dns-666b6646f7-sxgbr\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.793603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-config\") pod \"dnsmasq-dns-666b6646f7-sxgbr\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.794561 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-sxgbr\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.863895 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd4hb\" (UniqueName: \"kubernetes.io/projected/7456224a-635c-4191-ad12-e7ecd14192e6-kube-api-access-cd4hb\") pod \"dnsmasq-dns-666b6646f7-sxgbr\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.901005 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vmsqq"] Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.935324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.940069 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmzpm"] Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.941249 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:56 crc kubenswrapper[4824]: I1124 13:31:56.956192 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmzpm"] Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.098226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bmzpm\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.098290 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsk4s\" (UniqueName: \"kubernetes.io/projected/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-kube-api-access-zsk4s\") pod \"dnsmasq-dns-57d769cc4f-bmzpm\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.098488 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-config\") pod \"dnsmasq-dns-57d769cc4f-bmzpm\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.199446 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bmzpm\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.199490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsk4s\" (UniqueName: \"kubernetes.io/projected/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-kube-api-access-zsk4s\") pod \"dnsmasq-dns-57d769cc4f-bmzpm\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.199558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-config\") pod \"dnsmasq-dns-57d769cc4f-bmzpm\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.200404 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bmzpm\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.200942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-config\") pod \"dnsmasq-dns-57d769cc4f-bmzpm\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.236150 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsk4s\" (UniqueName: \"kubernetes.io/projected/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-kube-api-access-zsk4s\") pod \"dnsmasq-dns-57d769cc4f-bmzpm\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.262024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.722866 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sxgbr"] Nov 24 13:31:57 crc kubenswrapper[4824]: W1124 13:31:57.734520 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7456224a_635c_4191_ad12_e7ecd14192e6.slice/crio-ab2ef0f1ab29c68cd36916b13feacae7cf141f11723acfc0c0a747b537c3b61d WatchSource:0}: Error finding container ab2ef0f1ab29c68cd36916b13feacae7cf141f11723acfc0c0a747b537c3b61d: Status 404 returned error can't find the container with id ab2ef0f1ab29c68cd36916b13feacae7cf141f11723acfc0c0a747b537c3b61d Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.772530 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.773995 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.776510 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.781603 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.782214 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.782613 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.782787 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mfcwz" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.782791 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.783989 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.798833 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.819638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" event={"ID":"7456224a-635c-4191-ad12-e7ecd14192e6","Type":"ContainerStarted","Data":"ab2ef0f1ab29c68cd36916b13feacae7cf141f11723acfc0c0a747b537c3b61d"} Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.893943 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmzpm"] Nov 24 13:31:57 crc kubenswrapper[4824]: W1124 13:31:57.913938 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8df76b6_2771_4b24_84ef_ad8ee0ac4fed.slice/crio-e5330bcb251e33a6f15bbc036d34694b22d41f9738afdd110a62e36e6d8d1b7a WatchSource:0}: Error finding container e5330bcb251e33a6f15bbc036d34694b22d41f9738afdd110a62e36e6d8d1b7a: Status 404 returned error can't find the container with id e5330bcb251e33a6f15bbc036d34694b22d41f9738afdd110a62e36e6d8d1b7a Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921186 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-config-data\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ds8w\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-kube-api-access-2ds8w\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921368 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:57 crc kubenswrapper[4824]: I1124 13:31:57.921422 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.022650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ds8w\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-kube-api-access-2ds8w\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.022990 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.023018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.023038 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.030459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.030624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.030660 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.030861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.031075 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.031110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.031428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.031451 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.031120 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.031738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-config-data\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.032223 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.032379 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.033256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-config-data\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.034787 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.041590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.041824 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.044954 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.047664 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ds8w\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-kube-api-access-2ds8w\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.063114 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.078770 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.080320 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.093007 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.093454 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.093545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.093745 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.094155 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.097059 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.105561 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4xrjn" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.107195 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.123897 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245359 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245651 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245670 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245728 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc596e02-b760-49d3-a0c2-00ec8e751e05-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245827 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb4jb\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-kube-api-access-jb4jb\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245881 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.245901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc596e02-b760-49d3-a0c2-00ec8e751e05-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347116 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc596e02-b760-49d3-a0c2-00ec8e751e05-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347150 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347186 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347201 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb4jb\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-kube-api-access-jb4jb\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc596e02-b760-49d3-a0c2-00ec8e751e05-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347281 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.347324 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.348328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.349079 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.349768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.350172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.350373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.354379 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.357489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.370314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc596e02-b760-49d3-a0c2-00ec8e751e05-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.370370 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc596e02-b760-49d3-a0c2-00ec8e751e05-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.378000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb4jb\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-kube-api-access-jb4jb\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.421045 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.601209 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:31:58 crc kubenswrapper[4824]: W1124 13:31:58.616600 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f48dd6f_cca1_42fc_b4bd_bd841b8fb6c7.slice/crio-dfeb12863341e17eefcdd077ee34f2e1a72a4defa9795a5511355e904d743376 WatchSource:0}: Error finding container dfeb12863341e17eefcdd077ee34f2e1a72a4defa9795a5511355e904d743376: Status 404 returned error can't find the container with id dfeb12863341e17eefcdd077ee34f2e1a72a4defa9795a5511355e904d743376 Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.721886 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.874772 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7","Type":"ContainerStarted","Data":"dfeb12863341e17eefcdd077ee34f2e1a72a4defa9795a5511355e904d743376"} Nov 24 13:31:58 crc kubenswrapper[4824]: I1124 13:31:58.887999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" event={"ID":"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed","Type":"ContainerStarted","Data":"e5330bcb251e33a6f15bbc036d34694b22d41f9738afdd110a62e36e6d8d1b7a"} Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.353680 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:31:59 crc kubenswrapper[4824]: W1124 13:31:59.367887 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc596e02_b760_49d3_a0c2_00ec8e751e05.slice/crio-2982de524dd519d04b87f72eb9b3808a8c6550b9a6b1a5bdb14d0f2158f18dd8 WatchSource:0}: Error finding container 2982de524dd519d04b87f72eb9b3808a8c6550b9a6b1a5bdb14d0f2158f18dd8: Status 404 returned error can't find the container with id 2982de524dd519d04b87f72eb9b3808a8c6550b9a6b1a5bdb14d0f2158f18dd8 Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.521517 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.523016 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.530039 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-44mwh" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.530613 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.531228 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.533032 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.539245 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.577558 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.678398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.678481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e8da990-0e2a-47fd-b7d1-dec77870b40c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.678513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1e8da990-0e2a-47fd-b7d1-dec77870b40c-config-data-default\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.678538 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1e8da990-0e2a-47fd-b7d1-dec77870b40c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.678576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e8da990-0e2a-47fd-b7d1-dec77870b40c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.678728 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqxjv\" (UniqueName: \"kubernetes.io/projected/1e8da990-0e2a-47fd-b7d1-dec77870b40c-kube-api-access-pqxjv\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.678761 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1e8da990-0e2a-47fd-b7d1-dec77870b40c-kolla-config\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.678859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e8da990-0e2a-47fd-b7d1-dec77870b40c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.780695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqxjv\" (UniqueName: \"kubernetes.io/projected/1e8da990-0e2a-47fd-b7d1-dec77870b40c-kube-api-access-pqxjv\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.780746 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1e8da990-0e2a-47fd-b7d1-dec77870b40c-kolla-config\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.780772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e8da990-0e2a-47fd-b7d1-dec77870b40c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.780799 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.780870 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e8da990-0e2a-47fd-b7d1-dec77870b40c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.780898 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1e8da990-0e2a-47fd-b7d1-dec77870b40c-config-data-default\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.780927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1e8da990-0e2a-47fd-b7d1-dec77870b40c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.780980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e8da990-0e2a-47fd-b7d1-dec77870b40c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.783797 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.785673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1e8da990-0e2a-47fd-b7d1-dec77870b40c-config-data-default\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.787760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1e8da990-0e2a-47fd-b7d1-dec77870b40c-kolla-config\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.788027 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e8da990-0e2a-47fd-b7d1-dec77870b40c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.788357 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1e8da990-0e2a-47fd-b7d1-dec77870b40c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.788640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e8da990-0e2a-47fd-b7d1-dec77870b40c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.805937 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e8da990-0e2a-47fd-b7d1-dec77870b40c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.814122 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.835374 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqxjv\" (UniqueName: \"kubernetes.io/projected/1e8da990-0e2a-47fd-b7d1-dec77870b40c-kube-api-access-pqxjv\") pod \"openstack-galera-0\" (UID: \"1e8da990-0e2a-47fd-b7d1-dec77870b40c\") " pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.885708 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 13:31:59 crc kubenswrapper[4824]: I1124 13:31:59.903658 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bc596e02-b760-49d3-a0c2-00ec8e751e05","Type":"ContainerStarted","Data":"2982de524dd519d04b87f72eb9b3808a8c6550b9a6b1a5bdb14d0f2158f18dd8"} Nov 24 13:32:00 crc kubenswrapper[4824]: I1124 13:32:00.646928 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 13:32:00 crc kubenswrapper[4824]: W1124 13:32:00.726913 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e8da990_0e2a_47fd_b7d1_dec77870b40c.slice/crio-5a213caa5f08460753be4a35f72738668fe1e613a3a327c14549e600b74f0246 WatchSource:0}: Error finding container 5a213caa5f08460753be4a35f72738668fe1e613a3a327c14549e600b74f0246: Status 404 returned error can't find the container with id 5a213caa5f08460753be4a35f72738668fe1e613a3a327c14549e600b74f0246 Nov 24 13:32:00 crc kubenswrapper[4824]: I1124 13:32:00.961576 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 13:32:00 crc kubenswrapper[4824]: I1124 13:32:00.976380 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:00 crc kubenswrapper[4824]: I1124 13:32:00.980490 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-q54fh" Nov 24 13:32:00 crc kubenswrapper[4824]: I1124 13:32:00.984352 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 13:32:00 crc kubenswrapper[4824]: I1124 13:32:00.984788 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 13:32:00 crc kubenswrapper[4824]: I1124 13:32:00.987107 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 13:32:00 crc kubenswrapper[4824]: I1124 13:32:00.995747 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.079899 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1e8da990-0e2a-47fd-b7d1-dec77870b40c","Type":"ContainerStarted","Data":"5a213caa5f08460753be4a35f72738668fe1e613a3a327c14549e600b74f0246"} Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.116548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.116594 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90325e3b-c1cb-4244-a81b-fa666652304b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.116640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90325e3b-c1cb-4244-a81b-fa666652304b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.116658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90325e3b-c1cb-4244-a81b-fa666652304b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.116701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90325e3b-c1cb-4244-a81b-fa666652304b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.116720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90325e3b-c1cb-4244-a81b-fa666652304b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.116755 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/90325e3b-c1cb-4244-a81b-fa666652304b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.116817 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99pxn\" (UniqueName: \"kubernetes.io/projected/90325e3b-c1cb-4244-a81b-fa666652304b-kube-api-access-99pxn\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.147217 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.148225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.152922 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.163030 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.163371 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.163519 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-tzb74" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.219959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90325e3b-c1cb-4244-a81b-fa666652304b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220023 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/90325e3b-c1cb-4244-a81b-fa666652304b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220052 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99pxn\" (UniqueName: \"kubernetes.io/projected/90325e3b-c1cb-4244-a81b-fa666652304b-kube-api-access-99pxn\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220121 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90325e3b-c1cb-4244-a81b-fa666652304b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7687928-0bb1-456c-8555-e918db203197-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220164 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7687928-0bb1-456c-8555-e918db203197-kolla-config\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90325e3b-c1cb-4244-a81b-fa666652304b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220208 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90325e3b-c1cb-4244-a81b-fa666652304b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7687928-0bb1-456c-8555-e918db203197-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7687928-0bb1-456c-8555-e918db203197-config-data\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220274 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg58w\" (UniqueName: \"kubernetes.io/projected/a7687928-0bb1-456c-8555-e918db203197-kube-api-access-jg58w\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220291 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90325e3b-c1cb-4244-a81b-fa666652304b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90325e3b-c1cb-4244-a81b-fa666652304b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.220773 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.221657 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90325e3b-c1cb-4244-a81b-fa666652304b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.222000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90325e3b-c1cb-4244-a81b-fa666652304b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.222324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90325e3b-c1cb-4244-a81b-fa666652304b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.242830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/90325e3b-c1cb-4244-a81b-fa666652304b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.246380 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90325e3b-c1cb-4244-a81b-fa666652304b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.259098 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99pxn\" (UniqueName: \"kubernetes.io/projected/90325e3b-c1cb-4244-a81b-fa666652304b-kube-api-access-99pxn\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.300159 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"90325e3b-c1cb-4244-a81b-fa666652304b\") " pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.321550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7687928-0bb1-456c-8555-e918db203197-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.321595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7687928-0bb1-456c-8555-e918db203197-kolla-config\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.321636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7687928-0bb1-456c-8555-e918db203197-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.321658 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7687928-0bb1-456c-8555-e918db203197-config-data\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.321693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg58w\" (UniqueName: \"kubernetes.io/projected/a7687928-0bb1-456c-8555-e918db203197-kube-api-access-jg58w\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.322935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7687928-0bb1-456c-8555-e918db203197-kolla-config\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.325142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7687928-0bb1-456c-8555-e918db203197-config-data\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.332016 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.332521 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7687928-0bb1-456c-8555-e918db203197-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.334218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7687928-0bb1-456c-8555-e918db203197-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.364203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg58w\" (UniqueName: \"kubernetes.io/projected/a7687928-0bb1-456c-8555-e918db203197-kube-api-access-jg58w\") pod \"memcached-0\" (UID: \"a7687928-0bb1-456c-8555-e918db203197\") " pod="openstack/memcached-0" Nov 24 13:32:01 crc kubenswrapper[4824]: I1124 13:32:01.502279 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 13:32:02 crc kubenswrapper[4824]: I1124 13:32:02.048861 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 13:32:02 crc kubenswrapper[4824]: I1124 13:32:02.203751 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 13:32:02 crc kubenswrapper[4824]: I1124 13:32:02.731542 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:32:02 crc kubenswrapper[4824]: I1124 13:32:02.732484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:32:02 crc kubenswrapper[4824]: I1124 13:32:02.738826 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-ppmfv" Nov 24 13:32:02 crc kubenswrapper[4824]: I1124 13:32:02.758893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:32:02 crc kubenswrapper[4824]: I1124 13:32:02.860243 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmr6v\" (UniqueName: \"kubernetes.io/projected/25fd29f3-f0aa-4062-9f63-19fe3218583c-kube-api-access-cmr6v\") pod \"kube-state-metrics-0\" (UID: \"25fd29f3-f0aa-4062-9f63-19fe3218583c\") " pod="openstack/kube-state-metrics-0" Nov 24 13:32:02 crc kubenswrapper[4824]: I1124 13:32:02.961816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmr6v\" (UniqueName: \"kubernetes.io/projected/25fd29f3-f0aa-4062-9f63-19fe3218583c-kube-api-access-cmr6v\") pod \"kube-state-metrics-0\" (UID: \"25fd29f3-f0aa-4062-9f63-19fe3218583c\") " pod="openstack/kube-state-metrics-0" Nov 24 13:32:03 crc kubenswrapper[4824]: I1124 13:32:03.009395 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmr6v\" (UniqueName: \"kubernetes.io/projected/25fd29f3-f0aa-4062-9f63-19fe3218583c-kube-api-access-cmr6v\") pod \"kube-state-metrics-0\" (UID: \"25fd29f3-f0aa-4062-9f63-19fe3218583c\") " pod="openstack/kube-state-metrics-0" Nov 24 13:32:03 crc kubenswrapper[4824]: I1124 13:32:03.083315 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.299232 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.300963 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.305105 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-d64tx" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.314984 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.315601 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.315660 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.315779 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.317768 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.328284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h78hd\" (UniqueName: \"kubernetes.io/projected/7c045328-9e32-430c-b29b-91a3aeae62c7-kube-api-access-h78hd\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.328338 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.328379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c045328-9e32-430c-b29b-91a3aeae62c7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.328406 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c045328-9e32-430c-b29b-91a3aeae62c7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.328439 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c045328-9e32-430c-b29b-91a3aeae62c7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.328464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c045328-9e32-430c-b29b-91a3aeae62c7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.328519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c045328-9e32-430c-b29b-91a3aeae62c7-config\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.328551 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c045328-9e32-430c-b29b-91a3aeae62c7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.430771 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c045328-9e32-430c-b29b-91a3aeae62c7-config\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.430900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c045328-9e32-430c-b29b-91a3aeae62c7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.430944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h78hd\" (UniqueName: \"kubernetes.io/projected/7c045328-9e32-430c-b29b-91a3aeae62c7-kube-api-access-h78hd\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.430973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.431008 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c045328-9e32-430c-b29b-91a3aeae62c7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.431028 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c045328-9e32-430c-b29b-91a3aeae62c7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.431053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c045328-9e32-430c-b29b-91a3aeae62c7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.431070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c045328-9e32-430c-b29b-91a3aeae62c7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.431532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c045328-9e32-430c-b29b-91a3aeae62c7-config\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.431897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c045328-9e32-430c-b29b-91a3aeae62c7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.432557 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.434268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c045328-9e32-430c-b29b-91a3aeae62c7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.443672 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c045328-9e32-430c-b29b-91a3aeae62c7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.447620 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c045328-9e32-430c-b29b-91a3aeae62c7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.450938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c045328-9e32-430c-b29b-91a3aeae62c7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.451008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h78hd\" (UniqueName: \"kubernetes.io/projected/7c045328-9e32-430c-b29b-91a3aeae62c7-kube-api-access-h78hd\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.457515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7c045328-9e32-430c-b29b-91a3aeae62c7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.648634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.807742 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ds8fs"] Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.809172 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.816091 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-n6lfd" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.816138 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.828463 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.871549 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ds8fs"] Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.881640 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-jvdgl"] Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.904346 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.915761 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jvdgl"] Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.940668 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppq6r\" (UniqueName: \"kubernetes.io/projected/b3c188e2-9a89-47fe-8efe-1187a3c81552-kube-api-access-ppq6r\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.940952 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b3c188e2-9a89-47fe-8efe-1187a3c81552-var-run-ovn\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.941064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3c188e2-9a89-47fe-8efe-1187a3c81552-ovn-controller-tls-certs\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.941135 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c188e2-9a89-47fe-8efe-1187a3c81552-combined-ca-bundle\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.941217 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b3c188e2-9a89-47fe-8efe-1187a3c81552-var-run\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.941309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3c188e2-9a89-47fe-8efe-1187a3c81552-scripts\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:06 crc kubenswrapper[4824]: I1124 13:32:06.941386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b3c188e2-9a89-47fe-8efe-1187a3c81552-var-log-ovn\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045534 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-var-run\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-var-lib\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045619 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-etc-ovs\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045642 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3c188e2-9a89-47fe-8efe-1187a3c81552-ovn-controller-tls-certs\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045660 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c188e2-9a89-47fe-8efe-1187a3c81552-combined-ca-bundle\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b3c188e2-9a89-47fe-8efe-1187a3c81552-var-run\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045720 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3c188e2-9a89-47fe-8efe-1187a3c81552-scripts\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-var-log\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045753 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25645cc4-c970-453f-a6c3-08545ce4c84c-scripts\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045769 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b3c188e2-9a89-47fe-8efe-1187a3c81552-var-log-ovn\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045836 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qm2q\" (UniqueName: \"kubernetes.io/projected/25645cc4-c970-453f-a6c3-08545ce4c84c-kube-api-access-2qm2q\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045864 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppq6r\" (UniqueName: \"kubernetes.io/projected/b3c188e2-9a89-47fe-8efe-1187a3c81552-kube-api-access-ppq6r\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.045880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b3c188e2-9a89-47fe-8efe-1187a3c81552-var-run-ovn\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.046435 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b3c188e2-9a89-47fe-8efe-1187a3c81552-var-run-ovn\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.046558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b3c188e2-9a89-47fe-8efe-1187a3c81552-var-log-ovn\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.046912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b3c188e2-9a89-47fe-8efe-1187a3c81552-var-run\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.049090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3c188e2-9a89-47fe-8efe-1187a3c81552-scripts\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.051887 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3c188e2-9a89-47fe-8efe-1187a3c81552-ovn-controller-tls-certs\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.053401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c188e2-9a89-47fe-8efe-1187a3c81552-combined-ca-bundle\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.069438 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppq6r\" (UniqueName: \"kubernetes.io/projected/b3c188e2-9a89-47fe-8efe-1187a3c81552-kube-api-access-ppq6r\") pod \"ovn-controller-ds8fs\" (UID: \"b3c188e2-9a89-47fe-8efe-1187a3c81552\") " pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.144633 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.154142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-var-log\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.154224 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25645cc4-c970-453f-a6c3-08545ce4c84c-scripts\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.154307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qm2q\" (UniqueName: \"kubernetes.io/projected/25645cc4-c970-453f-a6c3-08545ce4c84c-kube-api-access-2qm2q\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.154396 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-var-log\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.154424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-var-run\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.156395 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-var-run\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.156403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25645cc4-c970-453f-a6c3-08545ce4c84c-scripts\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.157966 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-var-lib\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.158076 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-etc-ovs\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.158497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-var-lib\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.158499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/25645cc4-c970-453f-a6c3-08545ce4c84c-etc-ovs\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.170880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qm2q\" (UniqueName: \"kubernetes.io/projected/25645cc4-c970-453f-a6c3-08545ce4c84c-kube-api-access-2qm2q\") pod \"ovn-controller-ovs-jvdgl\" (UID: \"25645cc4-c970-453f-a6c3-08545ce4c84c\") " pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: I1124 13:32:07.222573 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:07 crc kubenswrapper[4824]: W1124 13:32:07.816960 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90325e3b_c1cb_4244_a81b_fa666652304b.slice/crio-9319420edc28fd05c36d231e88a01de43059b4d8398e4f8a435b92baa4123928 WatchSource:0}: Error finding container 9319420edc28fd05c36d231e88a01de43059b4d8398e4f8a435b92baa4123928: Status 404 returned error can't find the container with id 9319420edc28fd05c36d231e88a01de43059b4d8398e4f8a435b92baa4123928 Nov 24 13:32:08 crc kubenswrapper[4824]: I1124 13:32:08.144007 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a7687928-0bb1-456c-8555-e918db203197","Type":"ContainerStarted","Data":"d5e5c51b276c1cfd3d088ee69c4fa1bb1b68c8b079339f0fe3950bedaf2f5fda"} Nov 24 13:32:08 crc kubenswrapper[4824]: I1124 13:32:08.145269 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"90325e3b-c1cb-4244-a81b-fa666652304b","Type":"ContainerStarted","Data":"9319420edc28fd05c36d231e88a01de43059b4d8398e4f8a435b92baa4123928"} Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.433034 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.437237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.439626 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.439675 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.454626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.454911 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-fdsrp" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.457423 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.522952 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.523357 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.523468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.523496 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.523525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-config\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.523691 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.523757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8kgg\" (UniqueName: \"kubernetes.io/projected/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-kube-api-access-r8kgg\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.523851 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.624954 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.625001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.625024 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-config\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.625188 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.625214 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8kgg\" (UniqueName: \"kubernetes.io/projected/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-kube-api-access-r8kgg\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.625251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.625266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.625297 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.625633 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.625648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.627588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.643674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.644282 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.646576 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.648605 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8kgg\" (UniqueName: \"kubernetes.io/projected/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-kube-api-access-r8kgg\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.649447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11c797e-a7a6-4b6f-8261-cdb6bdc7a525-config\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.664739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525\") " pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.755289 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.787728 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.787789 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.787905 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.788644 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"62c6a3d914058b43356f45839d76c3c6a0c31d95c968d9b2fa78dd4dbda74ea8"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:32:10 crc kubenswrapper[4824]: I1124 13:32:10.788693 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://62c6a3d914058b43356f45839d76c3c6a0c31d95c968d9b2fa78dd4dbda74ea8" gracePeriod=600 Nov 24 13:32:11 crc kubenswrapper[4824]: I1124 13:32:11.169113 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="62c6a3d914058b43356f45839d76c3c6a0c31d95c968d9b2fa78dd4dbda74ea8" exitCode=0 Nov 24 13:32:11 crc kubenswrapper[4824]: I1124 13:32:11.169151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"62c6a3d914058b43356f45839d76c3c6a0c31d95c968d9b2fa78dd4dbda74ea8"} Nov 24 13:32:11 crc kubenswrapper[4824]: I1124 13:32:11.169179 4824 scope.go:117] "RemoveContainer" containerID="e56c6ac43b570680166358738b704b70ae28054c60648850ac0d5d06e92a00bf" Nov 24 13:32:28 crc kubenswrapper[4824]: E1124 13:32:28.136678 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Nov 24 13:32:28 crc kubenswrapper[4824]: E1124 13:32:28.138201 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n65dh77h649h547h5b5h678hcch5fbh5cch55dh66hcbh59ch84h678h67bhb4h54h54bh5cchfdh664h68bh65h5f8h96hfh87hf7h599h59bh577q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jg58w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(a7687928-0bb1-456c-8555-e918db203197): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:28 crc kubenswrapper[4824]: E1124 13:32:28.141030 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="a7687928-0bb1-456c-8555-e918db203197" Nov 24 13:32:28 crc kubenswrapper[4824]: E1124 13:32:28.331696 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="a7687928-0bb1-456c-8555-e918db203197" Nov 24 13:32:28 crc kubenswrapper[4824]: I1124 13:32:28.785413 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 13:32:30 crc kubenswrapper[4824]: W1124 13:32:30.055329 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c045328_9e32_430c_b29b_91a3aeae62c7.slice/crio-d267d5eb370cecf372ac2f671aae279165a0c33cf1a39a3c47bcee4d76f201c1 WatchSource:0}: Error finding container d267d5eb370cecf372ac2f671aae279165a0c33cf1a39a3c47bcee4d76f201c1: Status 404 returned error can't find the container with id d267d5eb370cecf372ac2f671aae279165a0c33cf1a39a3c47bcee4d76f201c1 Nov 24 13:32:30 crc kubenswrapper[4824]: I1124 13:32:30.312799 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7c045328-9e32-430c-b29b-91a3aeae62c7","Type":"ContainerStarted","Data":"d267d5eb370cecf372ac2f671aae279165a0c33cf1a39a3c47bcee4d76f201c1"} Nov 24 13:32:30 crc kubenswrapper[4824]: E1124 13:32:30.520236 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 13:32:30 crc kubenswrapper[4824]: E1124 13:32:30.520696 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zd8ww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-vmsqq_openstack(041daaa0-0f4f-45aa-bda7-589182e65764): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:30 crc kubenswrapper[4824]: E1124 13:32:30.523055 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" podUID="041daaa0-0f4f-45aa-bda7-589182e65764" Nov 24 13:32:30 crc kubenswrapper[4824]: E1124 13:32:30.846124 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 13:32:30 crc kubenswrapper[4824]: E1124 13:32:30.846283 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zsk4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-bmzpm_openstack(d8df76b6-2771-4b24-84ef-ad8ee0ac4fed): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:30 crc kubenswrapper[4824]: E1124 13:32:30.848023 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" podUID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" Nov 24 13:32:30 crc kubenswrapper[4824]: E1124 13:32:30.904643 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 13:32:30 crc kubenswrapper[4824]: E1124 13:32:30.905070 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p769b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-895h4_openstack(9f81886c-379a-4e1d-a5bb-961cbf65927c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:30 crc kubenswrapper[4824]: E1124 13:32:30.906408 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" podUID="9f81886c-379a-4e1d-a5bb-961cbf65927c" Nov 24 13:32:31 crc kubenswrapper[4824]: E1124 13:32:31.082070 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 13:32:31 crc kubenswrapper[4824]: E1124 13:32:31.082218 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cd4hb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-sxgbr_openstack(7456224a-635c-4191-ad12-e7ecd14192e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:32:31 crc kubenswrapper[4824]: E1124 13:32:31.083641 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" podUID="7456224a-635c-4191-ad12-e7ecd14192e6" Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.322867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"90325e3b-c1cb-4244-a81b-fa666652304b","Type":"ContainerStarted","Data":"3335dcc0a48e56590f22e53a3b44fb0121920db9360157d18b61cc5410fdcc7b"} Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.358002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"7023e2fedbc8a159e146d80c57d0a65acc7f7defe0350732f194124c1f095b6e"} Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.370713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1e8da990-0e2a-47fd-b7d1-dec77870b40c","Type":"ContainerStarted","Data":"e8df679820edbb8b1f9bae02d4f03b29cf7aed278d4afd9d7fe506c785f43126"} Nov 24 13:32:31 crc kubenswrapper[4824]: E1124 13:32:31.372246 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" podUID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" Nov 24 13:32:31 crc kubenswrapper[4824]: E1124 13:32:31.373502 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" podUID="7456224a-635c-4191-ad12-e7ecd14192e6" Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.422888 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ds8fs"] Nov 24 13:32:31 crc kubenswrapper[4824]: W1124 13:32:31.445344 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3c188e2_9a89_47fe_8efe_1187a3c81552.slice/crio-4c784202251e4c786c858285697c5a517bd78daea61ccb60b06389a1b31c9c38 WatchSource:0}: Error finding container 4c784202251e4c786c858285697c5a517bd78daea61ccb60b06389a1b31c9c38: Status 404 returned error can't find the container with id 4c784202251e4c786c858285697c5a517bd78daea61ccb60b06389a1b31c9c38 Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.559457 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.603433 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.676182 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jvdgl"] Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.765060 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.790151 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f81886c-379a-4e1d-a5bb-961cbf65927c-config\") pod \"9f81886c-379a-4e1d-a5bb-961cbf65927c\" (UID: \"9f81886c-379a-4e1d-a5bb-961cbf65927c\") " Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.790192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p769b\" (UniqueName: \"kubernetes.io/projected/9f81886c-379a-4e1d-a5bb-961cbf65927c-kube-api-access-p769b\") pod \"9f81886c-379a-4e1d-a5bb-961cbf65927c\" (UID: \"9f81886c-379a-4e1d-a5bb-961cbf65927c\") " Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.802766 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f81886c-379a-4e1d-a5bb-961cbf65927c-config" (OuterVolumeSpecName: "config") pod "9f81886c-379a-4e1d-a5bb-961cbf65927c" (UID: "9f81886c-379a-4e1d-a5bb-961cbf65927c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.805242 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f81886c-379a-4e1d-a5bb-961cbf65927c-kube-api-access-p769b" (OuterVolumeSpecName: "kube-api-access-p769b") pod "9f81886c-379a-4e1d-a5bb-961cbf65927c" (UID: "9f81886c-379a-4e1d-a5bb-961cbf65927c"). InnerVolumeSpecName "kube-api-access-p769b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.839968 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.891524 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f81886c-379a-4e1d-a5bb-961cbf65927c-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.891566 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p769b\" (UniqueName: \"kubernetes.io/projected/9f81886c-379a-4e1d-a5bb-961cbf65927c-kube-api-access-p769b\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.994627 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd8ww\" (UniqueName: \"kubernetes.io/projected/041daaa0-0f4f-45aa-bda7-589182e65764-kube-api-access-zd8ww\") pod \"041daaa0-0f4f-45aa-bda7-589182e65764\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.994768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-dns-svc\") pod \"041daaa0-0f4f-45aa-bda7-589182e65764\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.994876 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-config\") pod \"041daaa0-0f4f-45aa-bda7-589182e65764\" (UID: \"041daaa0-0f4f-45aa-bda7-589182e65764\") " Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.995939 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-config" (OuterVolumeSpecName: "config") pod "041daaa0-0f4f-45aa-bda7-589182e65764" (UID: "041daaa0-0f4f-45aa-bda7-589182e65764"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:31 crc kubenswrapper[4824]: I1124 13:32:31.996372 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "041daaa0-0f4f-45aa-bda7-589182e65764" (UID: "041daaa0-0f4f-45aa-bda7-589182e65764"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.079315 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/041daaa0-0f4f-45aa-bda7-589182e65764-kube-api-access-zd8ww" (OuterVolumeSpecName: "kube-api-access-zd8ww") pod "041daaa0-0f4f-45aa-bda7-589182e65764" (UID: "041daaa0-0f4f-45aa-bda7-589182e65764"). InnerVolumeSpecName "kube-api-access-zd8ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.096278 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.096315 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041daaa0-0f4f-45aa-bda7-589182e65764-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.096328 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd8ww\" (UniqueName: \"kubernetes.io/projected/041daaa0-0f4f-45aa-bda7-589182e65764-kube-api-access-zd8ww\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.393520 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7","Type":"ContainerStarted","Data":"3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c"} Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.400630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jvdgl" event={"ID":"25645cc4-c970-453f-a6c3-08545ce4c84c","Type":"ContainerStarted","Data":"9d0ce19d9c064c6f0a022b0a98c9396468838717437da7d42093566a31cf9e2a"} Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.403542 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.404183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-895h4" event={"ID":"9f81886c-379a-4e1d-a5bb-961cbf65927c","Type":"ContainerDied","Data":"42d413c2f9956dce31ad601e22f03c57fbd00f76628ed0eea6856604ba86291f"} Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.407628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25fd29f3-f0aa-4062-9f63-19fe3218583c","Type":"ContainerStarted","Data":"1f50a1af5f6c5e4f9a404b0f7360055587e390478b40d0e82a59ebc9c22a6f0e"} Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.411209 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bc596e02-b760-49d3-a0c2-00ec8e751e05","Type":"ContainerStarted","Data":"46d8ac6888d06952ac6295da8bf804f80e93e9aa76eb2e3004f7277c99076e12"} Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.426746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ds8fs" event={"ID":"b3c188e2-9a89-47fe-8efe-1187a3c81552","Type":"ContainerStarted","Data":"4c784202251e4c786c858285697c5a517bd78daea61ccb60b06389a1b31c9c38"} Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.433609 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" event={"ID":"041daaa0-0f4f-45aa-bda7-589182e65764","Type":"ContainerDied","Data":"aee4dd1676f49ad1367bf4664d47c4472f4d91674ad1d3c032ddda43ef2742bb"} Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.433697 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vmsqq" Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.460366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525","Type":"ContainerStarted","Data":"821a7653da19bfd234a484cfa59a7d20641885217a814a9651eb1bb8ec38902e"} Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.518873 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-895h4"] Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.522187 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-895h4"] Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.567809 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vmsqq"] Nov 24 13:32:32 crc kubenswrapper[4824]: I1124 13:32:32.590484 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vmsqq"] Nov 24 13:32:33 crc kubenswrapper[4824]: I1124 13:32:33.019998 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="041daaa0-0f4f-45aa-bda7-589182e65764" path="/var/lib/kubelet/pods/041daaa0-0f4f-45aa-bda7-589182e65764/volumes" Nov 24 13:32:33 crc kubenswrapper[4824]: I1124 13:32:33.020345 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f81886c-379a-4e1d-a5bb-961cbf65927c" path="/var/lib/kubelet/pods/9f81886c-379a-4e1d-a5bb-961cbf65927c/volumes" Nov 24 13:32:39 crc kubenswrapper[4824]: I1124 13:32:39.508632 4824 generic.go:334] "Generic (PLEG): container finished" podID="25645cc4-c970-453f-a6c3-08545ce4c84c" containerID="06b533730a1ca6cb999fd8edf7d2a8a3602a5c72b082c70d56a1d3d93f616eea" exitCode=0 Nov 24 13:32:39 crc kubenswrapper[4824]: I1124 13:32:39.509130 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jvdgl" event={"ID":"25645cc4-c970-453f-a6c3-08545ce4c84c","Type":"ContainerDied","Data":"06b533730a1ca6cb999fd8edf7d2a8a3602a5c72b082c70d56a1d3d93f616eea"} Nov 24 13:32:39 crc kubenswrapper[4824]: I1124 13:32:39.514535 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ds8fs" event={"ID":"b3c188e2-9a89-47fe-8efe-1187a3c81552","Type":"ContainerStarted","Data":"0223d7d80a92954b873466c79d1f986cab01944ca38e19b4c13894e33d89199c"} Nov 24 13:32:39 crc kubenswrapper[4824]: I1124 13:32:39.515066 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ds8fs" Nov 24 13:32:39 crc kubenswrapper[4824]: I1124 13:32:39.516430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525","Type":"ContainerStarted","Data":"047ac46e44e79e4ff73cb0900553b7ccafe443825ec59c6febe17ab37df686ae"} Nov 24 13:32:39 crc kubenswrapper[4824]: I1124 13:32:39.553161 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ds8fs" podStartSLOduration=26.836803625 podStartE2EDuration="33.553141624s" podCreationTimestamp="2025-11-24 13:32:06 +0000 UTC" firstStartedPulling="2025-11-24 13:32:31.451353767 +0000 UTC m=+1093.090893077" lastFinishedPulling="2025-11-24 13:32:38.167691766 +0000 UTC m=+1099.807231076" observedRunningTime="2025-11-24 13:32:39.545034456 +0000 UTC m=+1101.184573776" watchObservedRunningTime="2025-11-24 13:32:39.553141624 +0000 UTC m=+1101.192680944" Nov 24 13:32:40 crc kubenswrapper[4824]: I1124 13:32:40.526307 4824 generic.go:334] "Generic (PLEG): container finished" podID="1e8da990-0e2a-47fd-b7d1-dec77870b40c" containerID="e8df679820edbb8b1f9bae02d4f03b29cf7aed278d4afd9d7fe506c785f43126" exitCode=0 Nov 24 13:32:40 crc kubenswrapper[4824]: I1124 13:32:40.526387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1e8da990-0e2a-47fd-b7d1-dec77870b40c","Type":"ContainerDied","Data":"e8df679820edbb8b1f9bae02d4f03b29cf7aed278d4afd9d7fe506c785f43126"} Nov 24 13:32:40 crc kubenswrapper[4824]: I1124 13:32:40.533207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jvdgl" event={"ID":"25645cc4-c970-453f-a6c3-08545ce4c84c","Type":"ContainerStarted","Data":"51d4510f5549fdf5a9882d787adc50f06e5c79f6fe647e925e353ad0f618ec5c"} Nov 24 13:32:40 crc kubenswrapper[4824]: I1124 13:32:40.535906 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25fd29f3-f0aa-4062-9f63-19fe3218583c","Type":"ContainerStarted","Data":"e31406a580667045cf0323d2386cd7e782e8611712a0e39a616c2fb768c4114e"} Nov 24 13:32:40 crc kubenswrapper[4824]: I1124 13:32:40.536965 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 13:32:40 crc kubenswrapper[4824]: I1124 13:32:40.537706 4824 generic.go:334] "Generic (PLEG): container finished" podID="90325e3b-c1cb-4244-a81b-fa666652304b" containerID="3335dcc0a48e56590f22e53a3b44fb0121920db9360157d18b61cc5410fdcc7b" exitCode=0 Nov 24 13:32:40 crc kubenswrapper[4824]: I1124 13:32:40.537765 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"90325e3b-c1cb-4244-a81b-fa666652304b","Type":"ContainerDied","Data":"3335dcc0a48e56590f22e53a3b44fb0121920db9360157d18b61cc5410fdcc7b"} Nov 24 13:32:40 crc kubenswrapper[4824]: I1124 13:32:40.539006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7c045328-9e32-430c-b29b-91a3aeae62c7","Type":"ContainerStarted","Data":"acf44489d89f36f4ed7282c37c00a122e1d43084a956f232e5f43c801afb0613"} Nov 24 13:32:40 crc kubenswrapper[4824]: I1124 13:32:40.567432 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=30.33613836 podStartE2EDuration="38.567415561s" podCreationTimestamp="2025-11-24 13:32:02 +0000 UTC" firstStartedPulling="2025-11-24 13:32:31.698622794 +0000 UTC m=+1093.338162104" lastFinishedPulling="2025-11-24 13:32:39.929899995 +0000 UTC m=+1101.569439305" observedRunningTime="2025-11-24 13:32:40.567331899 +0000 UTC m=+1102.206871209" watchObservedRunningTime="2025-11-24 13:32:40.567415561 +0000 UTC m=+1102.206954861" Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.550702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jvdgl" event={"ID":"25645cc4-c970-453f-a6c3-08545ce4c84c","Type":"ContainerStarted","Data":"f278208a64c5ea5591924022c0f3081da4ab0100a17c6df3a1dcaaaf4da8b962"} Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.551330 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.567687 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"90325e3b-c1cb-4244-a81b-fa666652304b","Type":"ContainerStarted","Data":"3030d91abbb9275a99885c3e321f531a6782c865e68d95233a364d4632c6dbec"} Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.569413 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a7687928-0bb1-456c-8555-e918db203197","Type":"ContainerStarted","Data":"52295e0976337912131f58aa393a63791e08a7572b12a1698292766cc24832ef"} Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.570116 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.575521 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1e8da990-0e2a-47fd-b7d1-dec77870b40c","Type":"ContainerStarted","Data":"e5286638c6e5671ce33383602ee1228f414bee028ab31f1899f9df7d39356c6e"} Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.596193 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-jvdgl" podStartSLOduration=29.145016611 podStartE2EDuration="35.596178071s" podCreationTimestamp="2025-11-24 13:32:06 +0000 UTC" firstStartedPulling="2025-11-24 13:32:31.711767143 +0000 UTC m=+1093.351306443" lastFinishedPulling="2025-11-24 13:32:38.162928583 +0000 UTC m=+1099.802467903" observedRunningTime="2025-11-24 13:32:41.590505965 +0000 UTC m=+1103.230045275" watchObservedRunningTime="2025-11-24 13:32:41.596178071 +0000 UTC m=+1103.235717381" Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.613042 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=7.968331517 podStartE2EDuration="40.613017855s" podCreationTimestamp="2025-11-24 13:32:01 +0000 UTC" firstStartedPulling="2025-11-24 13:32:07.81192441 +0000 UTC m=+1069.451463720" lastFinishedPulling="2025-11-24 13:32:40.456610748 +0000 UTC m=+1102.096150058" observedRunningTime="2025-11-24 13:32:41.609633618 +0000 UTC m=+1103.249172938" watchObservedRunningTime="2025-11-24 13:32:41.613017855 +0000 UTC m=+1103.252557165" Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.636015 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.611549703 podStartE2EDuration="43.635993607s" podCreationTimestamp="2025-11-24 13:31:58 +0000 UTC" firstStartedPulling="2025-11-24 13:32:00.748964325 +0000 UTC m=+1062.388503635" lastFinishedPulling="2025-11-24 13:32:30.773408229 +0000 UTC m=+1092.412947539" observedRunningTime="2025-11-24 13:32:41.628798671 +0000 UTC m=+1103.268337981" watchObservedRunningTime="2025-11-24 13:32:41.635993607 +0000 UTC m=+1103.275532917" Nov 24 13:32:41 crc kubenswrapper[4824]: I1124 13:32:41.658002 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=19.700110696 podStartE2EDuration="42.657984763s" podCreationTimestamp="2025-11-24 13:31:59 +0000 UTC" firstStartedPulling="2025-11-24 13:32:07.818432507 +0000 UTC m=+1069.457971817" lastFinishedPulling="2025-11-24 13:32:30.776306574 +0000 UTC m=+1092.415845884" observedRunningTime="2025-11-24 13:32:41.649589077 +0000 UTC m=+1103.289128397" watchObservedRunningTime="2025-11-24 13:32:41.657984763 +0000 UTC m=+1103.297524073" Nov 24 13:32:42 crc kubenswrapper[4824]: I1124 13:32:42.223277 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:32:44 crc kubenswrapper[4824]: I1124 13:32:44.603394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7c045328-9e32-430c-b29b-91a3aeae62c7","Type":"ContainerStarted","Data":"412029eb980fd46a7c4c163ddc7f08713999a53af50d2c3e209fd14a69c81d46"} Nov 24 13:32:44 crc kubenswrapper[4824]: I1124 13:32:44.607744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c11c797e-a7a6-4b6f-8261-cdb6bdc7a525","Type":"ContainerStarted","Data":"9a42551186450871ceeb0d8c5778506fd0e647e894f36e76c4c47cfaf0baf1c4"} Nov 24 13:32:44 crc kubenswrapper[4824]: I1124 13:32:44.611614 4824 generic.go:334] "Generic (PLEG): container finished" podID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" containerID="2120295df108e1ce33721f7d4e42340723e328c3c70ea3615893e02c186ee9c8" exitCode=0 Nov 24 13:32:44 crc kubenswrapper[4824]: I1124 13:32:44.611682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" event={"ID":"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed","Type":"ContainerDied","Data":"2120295df108e1ce33721f7d4e42340723e328c3c70ea3615893e02c186ee9c8"} Nov 24 13:32:44 crc kubenswrapper[4824]: I1124 13:32:44.634864 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=26.299188177 podStartE2EDuration="39.634850118s" podCreationTimestamp="2025-11-24 13:32:05 +0000 UTC" firstStartedPulling="2025-11-24 13:32:30.460124392 +0000 UTC m=+1092.099663742" lastFinishedPulling="2025-11-24 13:32:43.795786373 +0000 UTC m=+1105.435325683" observedRunningTime="2025-11-24 13:32:44.632836136 +0000 UTC m=+1106.272375456" watchObservedRunningTime="2025-11-24 13:32:44.634850118 +0000 UTC m=+1106.274389418" Nov 24 13:32:44 crc kubenswrapper[4824]: I1124 13:32:44.660294 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=23.613477651 podStartE2EDuration="35.660277243s" podCreationTimestamp="2025-11-24 13:32:09 +0000 UTC" firstStartedPulling="2025-11-24 13:32:31.71203788 +0000 UTC m=+1093.351577180" lastFinishedPulling="2025-11-24 13:32:43.758837462 +0000 UTC m=+1105.398376772" observedRunningTime="2025-11-24 13:32:44.65236967 +0000 UTC m=+1106.291908980" watchObservedRunningTime="2025-11-24 13:32:44.660277243 +0000 UTC m=+1106.299816553" Nov 24 13:32:45 crc kubenswrapper[4824]: I1124 13:32:45.621492 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" event={"ID":"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed","Type":"ContainerStarted","Data":"5fd252434606f9e3d5b6b7edfc9438f6d5bc67565ffd566dfbb6b5ed1f217adb"} Nov 24 13:32:45 crc kubenswrapper[4824]: I1124 13:32:45.622574 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:32:45 crc kubenswrapper[4824]: I1124 13:32:45.642638 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" podStartSLOduration=3.763372904 podStartE2EDuration="49.642616859s" podCreationTimestamp="2025-11-24 13:31:56 +0000 UTC" firstStartedPulling="2025-11-24 13:31:57.918424136 +0000 UTC m=+1059.557963446" lastFinishedPulling="2025-11-24 13:32:43.797668091 +0000 UTC m=+1105.437207401" observedRunningTime="2025-11-24 13:32:45.638045801 +0000 UTC m=+1107.277585131" watchObservedRunningTime="2025-11-24 13:32:45.642616859 +0000 UTC m=+1107.282156179" Nov 24 13:32:45 crc kubenswrapper[4824]: I1124 13:32:45.649915 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:45 crc kubenswrapper[4824]: I1124 13:32:45.703083 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:45 crc kubenswrapper[4824]: I1124 13:32:45.755831 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:46 crc kubenswrapper[4824]: I1124 13:32:46.503517 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 13:32:46 crc kubenswrapper[4824]: I1124 13:32:46.633542 4824 generic.go:334] "Generic (PLEG): container finished" podID="7456224a-635c-4191-ad12-e7ecd14192e6" containerID="09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d" exitCode=0 Nov 24 13:32:46 crc kubenswrapper[4824]: I1124 13:32:46.633799 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" event={"ID":"7456224a-635c-4191-ad12-e7ecd14192e6","Type":"ContainerDied","Data":"09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d"} Nov 24 13:32:46 crc kubenswrapper[4824]: I1124 13:32:46.635504 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:46 crc kubenswrapper[4824]: I1124 13:32:46.758767 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:46 crc kubenswrapper[4824]: I1124 13:32:46.928186 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:46 crc kubenswrapper[4824]: I1124 13:32:46.960380 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.294843 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmzpm"] Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.322729 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2cfsp"] Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.324035 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.326341 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.330845 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2cfsp"] Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.459570 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-wb2pp"] Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.460744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.465681 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.470319 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-wb2pp"] Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.503650 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlm4m\" (UniqueName: \"kubernetes.io/projected/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-kube-api-access-zlm4m\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.503712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-config\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.503736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.503862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607674 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-ovn-rundir\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-config\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607772 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6t4s\" (UniqueName: \"kubernetes.io/projected/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-kube-api-access-b6t4s\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607791 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-combined-ca-bundle\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607850 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlm4m\" (UniqueName: \"kubernetes.io/projected/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-kube-api-access-zlm4m\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607867 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-config\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607882 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-ovs-rundir\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607931 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.607971 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.609162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.609259 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-config\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.610145 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.630011 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlm4m\" (UniqueName: \"kubernetes.io/projected/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-kube-api-access-zlm4m\") pod \"dnsmasq-dns-7fd796d7df-2cfsp\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.640049 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.640994 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" podUID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" containerName="dnsmasq-dns" containerID="cri-o://5fd252434606f9e3d5b6b7edfc9438f6d5bc67565ffd566dfbb6b5ed1f217adb" gracePeriod=10 Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.703102 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sxgbr"] Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.709669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-ovs-rundir\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.709766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.709832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-ovn-rundir\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.709884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-config\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.709920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6t4s\" (UniqueName: \"kubernetes.io/projected/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-kube-api-access-b6t4s\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.709943 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-combined-ca-bundle\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.710019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-ovs-rundir\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.710096 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-ovn-rundir\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.711193 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-config\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.717640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-combined-ca-bundle\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.721640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.731031 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.737969 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6t4s\" (UniqueName: \"kubernetes.io/projected/e586afb0-4f13-437a-b00f-ea7e1f6d82ad-kube-api-access-b6t4s\") pod \"ovn-controller-metrics-wb2pp\" (UID: \"e586afb0-4f13-437a-b00f-ea7e1f6d82ad\") " pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.746211 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qdtw7"] Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.747528 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.762177 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.777475 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qdtw7"] Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.784556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-wb2pp" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.916235 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nbqm\" (UniqueName: \"kubernetes.io/projected/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-kube-api-access-2nbqm\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.916621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-config\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.916646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.916675 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:47 crc kubenswrapper[4824]: I1124 13:32:47.916713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.006675 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.017750 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-config\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.017804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.017847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.017902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.017944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nbqm\" (UniqueName: \"kubernetes.io/projected/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-kube-api-access-2nbqm\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.018721 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-config\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.018765 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.019399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.021760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.023105 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.029548 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.039935 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.040111 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.040209 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.040316 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8fls5" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.056464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nbqm\" (UniqueName: \"kubernetes.io/projected/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-kube-api-access-2nbqm\") pod \"dnsmasq-dns-86db49b7ff-qdtw7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.110098 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2cfsp"] Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.121305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/102b3dcf-b337-4f91-954b-a4996ba67ca7-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.121358 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/102b3dcf-b337-4f91-954b-a4996ba67ca7-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.121382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhlvn\" (UniqueName: \"kubernetes.io/projected/102b3dcf-b337-4f91-954b-a4996ba67ca7-kube-api-access-mhlvn\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.121408 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102b3dcf-b337-4f91-954b-a4996ba67ca7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.121433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/102b3dcf-b337-4f91-954b-a4996ba67ca7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.121453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/102b3dcf-b337-4f91-954b-a4996ba67ca7-scripts\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.121466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/102b3dcf-b337-4f91-954b-a4996ba67ca7-config\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.159242 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.224308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/102b3dcf-b337-4f91-954b-a4996ba67ca7-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.224582 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/102b3dcf-b337-4f91-954b-a4996ba67ca7-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.224602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhlvn\" (UniqueName: \"kubernetes.io/projected/102b3dcf-b337-4f91-954b-a4996ba67ca7-kube-api-access-mhlvn\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.224629 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102b3dcf-b337-4f91-954b-a4996ba67ca7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.224656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/102b3dcf-b337-4f91-954b-a4996ba67ca7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.224675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/102b3dcf-b337-4f91-954b-a4996ba67ca7-scripts\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.224691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/102b3dcf-b337-4f91-954b-a4996ba67ca7-config\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.225567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/102b3dcf-b337-4f91-954b-a4996ba67ca7-config\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.227006 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/102b3dcf-b337-4f91-954b-a4996ba67ca7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.227488 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/102b3dcf-b337-4f91-954b-a4996ba67ca7-scripts\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.230400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/102b3dcf-b337-4f91-954b-a4996ba67ca7-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.230917 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/102b3dcf-b337-4f91-954b-a4996ba67ca7-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.244405 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhlvn\" (UniqueName: \"kubernetes.io/projected/102b3dcf-b337-4f91-954b-a4996ba67ca7-kube-api-access-mhlvn\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.253315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102b3dcf-b337-4f91-954b-a4996ba67ca7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"102b3dcf-b337-4f91-954b-a4996ba67ca7\") " pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.407244 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.456045 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-wb2pp"] Nov 24 13:32:48 crc kubenswrapper[4824]: W1124 13:32:48.469637 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode586afb0_4f13_437a_b00f_ea7e1f6d82ad.slice/crio-31883db6c6f25808cc365f35292305eaa291f07473ec1a9b37ba928fa51115f4 WatchSource:0}: Error finding container 31883db6c6f25808cc365f35292305eaa291f07473ec1a9b37ba928fa51115f4: Status 404 returned error can't find the container with id 31883db6c6f25808cc365f35292305eaa291f07473ec1a9b37ba928fa51115f4 Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.649682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" event={"ID":"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f","Type":"ContainerStarted","Data":"4f867d156b7f9c17122e455ceea67f0cb883d092774f988e7885993834927f9e"} Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.655559 4824 generic.go:334] "Generic (PLEG): container finished" podID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" containerID="5fd252434606f9e3d5b6b7edfc9438f6d5bc67565ffd566dfbb6b5ed1f217adb" exitCode=0 Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.655609 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" event={"ID":"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed","Type":"ContainerDied","Data":"5fd252434606f9e3d5b6b7edfc9438f6d5bc67565ffd566dfbb6b5ed1f217adb"} Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.658510 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-wb2pp" event={"ID":"e586afb0-4f13-437a-b00f-ea7e1f6d82ad","Type":"ContainerStarted","Data":"31883db6c6f25808cc365f35292305eaa291f07473ec1a9b37ba928fa51115f4"} Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.663045 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qdtw7"] Nov 24 13:32:48 crc kubenswrapper[4824]: W1124 13:32:48.667977 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8295b1a_dd8c_47f6_8fd3_5d7c36cfbfd7.slice/crio-e9c26b96f5c96bad83f9a2d310e4df743e1fe198f9df0dd76f7f81257199acae WatchSource:0}: Error finding container e9c26b96f5c96bad83f9a2d310e4df743e1fe198f9df0dd76f7f81257199acae: Status 404 returned error can't find the container with id e9c26b96f5c96bad83f9a2d310e4df743e1fe198f9df0dd76f7f81257199acae Nov 24 13:32:48 crc kubenswrapper[4824]: I1124 13:32:48.835169 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 13:32:48 crc kubenswrapper[4824]: W1124 13:32:48.848682 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod102b3dcf_b337_4f91_954b_a4996ba67ca7.slice/crio-c4512c827a4bb7bbae30abcd5c7843bb76edeb1c7810a11fe3594856b382475e WatchSource:0}: Error finding container c4512c827a4bb7bbae30abcd5c7843bb76edeb1c7810a11fe3594856b382475e: Status 404 returned error can't find the container with id c4512c827a4bb7bbae30abcd5c7843bb76edeb1c7810a11fe3594856b382475e Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.668310 4824 generic.go:334] "Generic (PLEG): container finished" podID="f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" containerID="605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8" exitCode=0 Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.668647 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" event={"ID":"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f","Type":"ContainerDied","Data":"605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8"} Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.674916 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-wb2pp" event={"ID":"e586afb0-4f13-437a-b00f-ea7e1f6d82ad","Type":"ContainerStarted","Data":"81bf1f45b3b66c7098a036daf86b3fdb77a56120664f24434e39f7eed468e3ec"} Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.676745 4824 generic.go:334] "Generic (PLEG): container finished" podID="e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" containerID="50f158d4f79ab9e8aeca8137010d3e8b30fc643cb0ac76010000a7c376828f75" exitCode=0 Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.676847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" event={"ID":"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7","Type":"ContainerDied","Data":"50f158d4f79ab9e8aeca8137010d3e8b30fc643cb0ac76010000a7c376828f75"} Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.676896 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" event={"ID":"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7","Type":"ContainerStarted","Data":"e9c26b96f5c96bad83f9a2d310e4df743e1fe198f9df0dd76f7f81257199acae"} Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.679886 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" event={"ID":"7456224a-635c-4191-ad12-e7ecd14192e6","Type":"ContainerStarted","Data":"bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b"} Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.680624 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.680836 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" podUID="7456224a-635c-4191-ad12-e7ecd14192e6" containerName="dnsmasq-dns" containerID="cri-o://bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b" gracePeriod=10 Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.683065 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"102b3dcf-b337-4f91-954b-a4996ba67ca7","Type":"ContainerStarted","Data":"c4512c827a4bb7bbae30abcd5c7843bb76edeb1c7810a11fe3594856b382475e"} Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.719593 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-wb2pp" podStartSLOduration=2.719576023 podStartE2EDuration="2.719576023s" podCreationTimestamp="2025-11-24 13:32:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:49.711212738 +0000 UTC m=+1111.350752068" watchObservedRunningTime="2025-11-24 13:32:49.719576023 +0000 UTC m=+1111.359115333" Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.763538 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" podStartSLOduration=-9223371983.09126 podStartE2EDuration="53.763515404s" podCreationTimestamp="2025-11-24 13:31:56 +0000 UTC" firstStartedPulling="2025-11-24 13:31:57.741770068 +0000 UTC m=+1059.381309378" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:49.733550353 +0000 UTC m=+1111.373089663" watchObservedRunningTime="2025-11-24 13:32:49.763515404 +0000 UTC m=+1111.403054714" Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.887128 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.887168 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 13:32:49 crc kubenswrapper[4824]: I1124 13:32:49.905075 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.098813 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-config\") pod \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.099067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsk4s\" (UniqueName: \"kubernetes.io/projected/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-kube-api-access-zsk4s\") pod \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.099122 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-dns-svc\") pod \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\" (UID: \"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed\") " Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.122474 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-kube-api-access-zsk4s" (OuterVolumeSpecName: "kube-api-access-zsk4s") pod "d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" (UID: "d8df76b6-2771-4b24-84ef-ad8ee0ac4fed"). InnerVolumeSpecName "kube-api-access-zsk4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.124513 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.201555 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsk4s\" (UniqueName: \"kubernetes.io/projected/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-kube-api-access-zsk4s\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.208858 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" (UID: "d8df76b6-2771-4b24-84ef-ad8ee0ac4fed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.243111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-config" (OuterVolumeSpecName: "config") pod "d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" (UID: "d8df76b6-2771-4b24-84ef-ad8ee0ac4fed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.276031 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.303044 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.303079 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.403902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd4hb\" (UniqueName: \"kubernetes.io/projected/7456224a-635c-4191-ad12-e7ecd14192e6-kube-api-access-cd4hb\") pod \"7456224a-635c-4191-ad12-e7ecd14192e6\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.404244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-dns-svc\") pod \"7456224a-635c-4191-ad12-e7ecd14192e6\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.404445 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-config\") pod \"7456224a-635c-4191-ad12-e7ecd14192e6\" (UID: \"7456224a-635c-4191-ad12-e7ecd14192e6\") " Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.413336 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7456224a-635c-4191-ad12-e7ecd14192e6-kube-api-access-cd4hb" (OuterVolumeSpecName: "kube-api-access-cd4hb") pod "7456224a-635c-4191-ad12-e7ecd14192e6" (UID: "7456224a-635c-4191-ad12-e7ecd14192e6"). InnerVolumeSpecName "kube-api-access-cd4hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.465838 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-config" (OuterVolumeSpecName: "config") pod "7456224a-635c-4191-ad12-e7ecd14192e6" (UID: "7456224a-635c-4191-ad12-e7ecd14192e6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.477243 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7456224a-635c-4191-ad12-e7ecd14192e6" (UID: "7456224a-635c-4191-ad12-e7ecd14192e6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.506503 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd4hb\" (UniqueName: \"kubernetes.io/projected/7456224a-635c-4191-ad12-e7ecd14192e6-kube-api-access-cd4hb\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.506544 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.506555 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7456224a-635c-4191-ad12-e7ecd14192e6-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.697627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" event={"ID":"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f","Type":"ContainerStarted","Data":"742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb"} Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.698077 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.704523 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" event={"ID":"d8df76b6-2771-4b24-84ef-ad8ee0ac4fed","Type":"ContainerDied","Data":"e5330bcb251e33a6f15bbc036d34694b22d41f9738afdd110a62e36e6d8d1b7a"} Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.704579 4824 scope.go:117] "RemoveContainer" containerID="5fd252434606f9e3d5b6b7edfc9438f6d5bc67565ffd566dfbb6b5ed1f217adb" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.704583 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmzpm" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.719477 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" podStartSLOduration=3.719458831 podStartE2EDuration="3.719458831s" podCreationTimestamp="2025-11-24 13:32:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:50.716950506 +0000 UTC m=+1112.356489816" watchObservedRunningTime="2025-11-24 13:32:50.719458831 +0000 UTC m=+1112.358998141" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.720542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" event={"ID":"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7","Type":"ContainerStarted","Data":"201ed3365f0b89394e4fa4fb5ecf036ea8cd108265816cac747fbd7a17870f85"} Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.720793 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.722886 4824 generic.go:334] "Generic (PLEG): container finished" podID="7456224a-635c-4191-ad12-e7ecd14192e6" containerID="bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b" exitCode=0 Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.723109 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" event={"ID":"7456224a-635c-4191-ad12-e7ecd14192e6","Type":"ContainerDied","Data":"bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b"} Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.723355 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" event={"ID":"7456224a-635c-4191-ad12-e7ecd14192e6","Type":"ContainerDied","Data":"ab2ef0f1ab29c68cd36916b13feacae7cf141f11723acfc0c0a747b537c3b61d"} Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.724082 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-sxgbr" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.731918 4824 scope.go:117] "RemoveContainer" containerID="2120295df108e1ce33721f7d4e42340723e328c3c70ea3615893e02c186ee9c8" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.745529 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" podStartSLOduration=3.745501501 podStartE2EDuration="3.745501501s" podCreationTimestamp="2025-11-24 13:32:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:50.742181136 +0000 UTC m=+1112.381720456" watchObservedRunningTime="2025-11-24 13:32:50.745501501 +0000 UTC m=+1112.385040811" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.767227 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmzpm"] Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.777142 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmzpm"] Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.787236 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sxgbr"] Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.792857 4824 scope.go:117] "RemoveContainer" containerID="bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.796205 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sxgbr"] Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.825402 4824 scope.go:117] "RemoveContainer" containerID="09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.861252 4824 scope.go:117] "RemoveContainer" containerID="bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.861621 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 13:32:50 crc kubenswrapper[4824]: E1124 13:32:50.862057 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b\": container with ID starting with bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b not found: ID does not exist" containerID="bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.862097 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b"} err="failed to get container status \"bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b\": rpc error: code = NotFound desc = could not find container \"bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b\": container with ID starting with bc508f78ebe8d85bbf0cb81fad00bef2c297ddbc34acc26cacb6b31acd4d3f2b not found: ID does not exist" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.862126 4824 scope.go:117] "RemoveContainer" containerID="09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d" Nov 24 13:32:50 crc kubenswrapper[4824]: E1124 13:32:50.862442 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d\": container with ID starting with 09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d not found: ID does not exist" containerID="09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d" Nov 24 13:32:50 crc kubenswrapper[4824]: I1124 13:32:50.862469 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d"} err="failed to get container status \"09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d\": rpc error: code = NotFound desc = could not find container \"09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d\": container with ID starting with 09c9184a9b2c9ca77cdaf00863e7532ff16553a13bf00582ca66005aae0d803d not found: ID does not exist" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.019332 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7456224a-635c-4191-ad12-e7ecd14192e6" path="/var/lib/kubelet/pods/7456224a-635c-4191-ad12-e7ecd14192e6/volumes" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.020145 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" path="/var/lib/kubelet/pods/d8df76b6-2771-4b24-84ef-ad8ee0ac4fed/volumes" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.056930 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-3410-account-create-c74k2"] Nov 24 13:32:51 crc kubenswrapper[4824]: E1124 13:32:51.057212 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7456224a-635c-4191-ad12-e7ecd14192e6" containerName="init" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.057228 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7456224a-635c-4191-ad12-e7ecd14192e6" containerName="init" Nov 24 13:32:51 crc kubenswrapper[4824]: E1124 13:32:51.057254 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" containerName="dnsmasq-dns" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.057260 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" containerName="dnsmasq-dns" Nov 24 13:32:51 crc kubenswrapper[4824]: E1124 13:32:51.057274 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7456224a-635c-4191-ad12-e7ecd14192e6" containerName="dnsmasq-dns" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.057280 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7456224a-635c-4191-ad12-e7ecd14192e6" containerName="dnsmasq-dns" Nov 24 13:32:51 crc kubenswrapper[4824]: E1124 13:32:51.057294 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" containerName="init" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.057299 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" containerName="init" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.057434 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8df76b6-2771-4b24-84ef-ad8ee0ac4fed" containerName="dnsmasq-dns" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.057454 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7456224a-635c-4191-ad12-e7ecd14192e6" containerName="dnsmasq-dns" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.057930 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.061148 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.073588 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3410-account-create-c74k2"] Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.142408 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2gs6v"] Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.145538 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.161902 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2gs6v"] Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.227970 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rclz\" (UniqueName: \"kubernetes.io/projected/827adb53-2cb6-46ec-be78-6d25ced71cff-kube-api-access-8rclz\") pod \"keystone-db-create-2gs6v\" (UID: \"827adb53-2cb6-46ec-be78-6d25ced71cff\") " pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.228034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp6jb\" (UniqueName: \"kubernetes.io/projected/01ac49a1-753f-46bf-bddd-410c28fe864e-kube-api-access-dp6jb\") pod \"keystone-3410-account-create-c74k2\" (UID: \"01ac49a1-753f-46bf-bddd-410c28fe864e\") " pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.228124 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/827adb53-2cb6-46ec-be78-6d25ced71cff-operator-scripts\") pod \"keystone-db-create-2gs6v\" (UID: \"827adb53-2cb6-46ec-be78-6d25ced71cff\") " pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.228146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ac49a1-753f-46bf-bddd-410c28fe864e-operator-scripts\") pod \"keystone-3410-account-create-c74k2\" (UID: \"01ac49a1-753f-46bf-bddd-410c28fe864e\") " pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.329574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ac49a1-753f-46bf-bddd-410c28fe864e-operator-scripts\") pod \"keystone-3410-account-create-c74k2\" (UID: \"01ac49a1-753f-46bf-bddd-410c28fe864e\") " pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.329666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rclz\" (UniqueName: \"kubernetes.io/projected/827adb53-2cb6-46ec-be78-6d25ced71cff-kube-api-access-8rclz\") pod \"keystone-db-create-2gs6v\" (UID: \"827adb53-2cb6-46ec-be78-6d25ced71cff\") " pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.329701 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp6jb\" (UniqueName: \"kubernetes.io/projected/01ac49a1-753f-46bf-bddd-410c28fe864e-kube-api-access-dp6jb\") pod \"keystone-3410-account-create-c74k2\" (UID: \"01ac49a1-753f-46bf-bddd-410c28fe864e\") " pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.329792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/827adb53-2cb6-46ec-be78-6d25ced71cff-operator-scripts\") pod \"keystone-db-create-2gs6v\" (UID: \"827adb53-2cb6-46ec-be78-6d25ced71cff\") " pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.330596 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/827adb53-2cb6-46ec-be78-6d25ced71cff-operator-scripts\") pod \"keystone-db-create-2gs6v\" (UID: \"827adb53-2cb6-46ec-be78-6d25ced71cff\") " pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.330632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ac49a1-753f-46bf-bddd-410c28fe864e-operator-scripts\") pod \"keystone-3410-account-create-c74k2\" (UID: \"01ac49a1-753f-46bf-bddd-410c28fe864e\") " pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.333422 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.333456 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.355995 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rclz\" (UniqueName: \"kubernetes.io/projected/827adb53-2cb6-46ec-be78-6d25ced71cff-kube-api-access-8rclz\") pod \"keystone-db-create-2gs6v\" (UID: \"827adb53-2cb6-46ec-be78-6d25ced71cff\") " pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.360483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp6jb\" (UniqueName: \"kubernetes.io/projected/01ac49a1-753f-46bf-bddd-410c28fe864e-kube-api-access-dp6jb\") pod \"keystone-3410-account-create-c74k2\" (UID: \"01ac49a1-753f-46bf-bddd-410c28fe864e\") " pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.365370 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-lsnff"] Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.367159 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lsnff" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.380596 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-lsnff"] Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.405140 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.431554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmq7h\" (UniqueName: \"kubernetes.io/projected/cbba9a43-445e-4657-9a7b-35cd2d2dc405-kube-api-access-dmq7h\") pod \"placement-db-create-lsnff\" (UID: \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\") " pod="openstack/placement-db-create-lsnff" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.431618 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba9a43-445e-4657-9a7b-35cd2d2dc405-operator-scripts\") pod \"placement-db-create-lsnff\" (UID: \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\") " pod="openstack/placement-db-create-lsnff" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.465299 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.487792 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-b3c8-account-create-n8x6v"] Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.489045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.501751 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b3c8-account-create-n8x6v"] Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.527028 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.527226 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.536643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmq7h\" (UniqueName: \"kubernetes.io/projected/cbba9a43-445e-4657-9a7b-35cd2d2dc405-kube-api-access-dmq7h\") pod \"placement-db-create-lsnff\" (UID: \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\") " pod="openstack/placement-db-create-lsnff" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.536683 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba9a43-445e-4657-9a7b-35cd2d2dc405-operator-scripts\") pod \"placement-db-create-lsnff\" (UID: \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\") " pod="openstack/placement-db-create-lsnff" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.539573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba9a43-445e-4657-9a7b-35cd2d2dc405-operator-scripts\") pod \"placement-db-create-lsnff\" (UID: \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\") " pod="openstack/placement-db-create-lsnff" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.579207 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmq7h\" (UniqueName: \"kubernetes.io/projected/cbba9a43-445e-4657-9a7b-35cd2d2dc405-kube-api-access-dmq7h\") pod \"placement-db-create-lsnff\" (UID: \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\") " pod="openstack/placement-db-create-lsnff" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.638406 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb47c\" (UniqueName: \"kubernetes.io/projected/e25c9bc5-d0fc-48e4-9579-59b360e49731-kube-api-access-wb47c\") pod \"placement-b3c8-account-create-n8x6v\" (UID: \"e25c9bc5-d0fc-48e4-9579-59b360e49731\") " pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.638689 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25c9bc5-d0fc-48e4-9579-59b360e49731-operator-scripts\") pod \"placement-b3c8-account-create-n8x6v\" (UID: \"e25c9bc5-d0fc-48e4-9579-59b360e49731\") " pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.726502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lsnff" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.739764 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb47c\" (UniqueName: \"kubernetes.io/projected/e25c9bc5-d0fc-48e4-9579-59b360e49731-kube-api-access-wb47c\") pod \"placement-b3c8-account-create-n8x6v\" (UID: \"e25c9bc5-d0fc-48e4-9579-59b360e49731\") " pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.739910 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25c9bc5-d0fc-48e4-9579-59b360e49731-operator-scripts\") pod \"placement-b3c8-account-create-n8x6v\" (UID: \"e25c9bc5-d0fc-48e4-9579-59b360e49731\") " pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.740693 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25c9bc5-d0fc-48e4-9579-59b360e49731-operator-scripts\") pod \"placement-b3c8-account-create-n8x6v\" (UID: \"e25c9bc5-d0fc-48e4-9579-59b360e49731\") " pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.746819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"102b3dcf-b337-4f91-954b-a4996ba67ca7","Type":"ContainerStarted","Data":"ff8b0ee3329208b72d1d4b14326237f548dfde9893f5b376cd6ad295a6a25656"} Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.746858 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"102b3dcf-b337-4f91-954b-a4996ba67ca7","Type":"ContainerStarted","Data":"648d5babdd264ec42582f27ffc2b38ef4e20587941ac762a6bc7ea9aee3741be"} Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.751269 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.776428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb47c\" (UniqueName: \"kubernetes.io/projected/e25c9bc5-d0fc-48e4-9579-59b360e49731-kube-api-access-wb47c\") pod \"placement-b3c8-account-create-n8x6v\" (UID: \"e25c9bc5-d0fc-48e4-9579-59b360e49731\") " pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.782145 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.936043746 podStartE2EDuration="4.782122344s" podCreationTimestamp="2025-11-24 13:32:47 +0000 UTC" firstStartedPulling="2025-11-24 13:32:48.850153235 +0000 UTC m=+1110.489692545" lastFinishedPulling="2025-11-24 13:32:50.696231833 +0000 UTC m=+1112.335771143" observedRunningTime="2025-11-24 13:32:51.776678854 +0000 UTC m=+1113.416218164" watchObservedRunningTime="2025-11-24 13:32:51.782122344 +0000 UTC m=+1113.421661654" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.859609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:51 crc kubenswrapper[4824]: I1124 13:32:51.888467 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.012653 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3410-account-create-c74k2"] Nov 24 13:32:52 crc kubenswrapper[4824]: W1124 13:32:52.030874 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ac49a1_753f_46bf_bddd_410c28fe864e.slice/crio-65fb31e35dc837c931c122ee18e449d452d8e62743e41017b02d5e871b264616 WatchSource:0}: Error finding container 65fb31e35dc837c931c122ee18e449d452d8e62743e41017b02d5e871b264616: Status 404 returned error can't find the container with id 65fb31e35dc837c931c122ee18e449d452d8e62743e41017b02d5e871b264616 Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.059646 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2gs6v"] Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.336475 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-lsnff"] Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.391378 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b3c8-account-create-n8x6v"] Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.772901 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lsnff" event={"ID":"cbba9a43-445e-4657-9a7b-35cd2d2dc405","Type":"ContainerStarted","Data":"987c50051bd0e9f80f7397d64a85e667b6a2c95451fc31064e48afcda69b206b"} Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.773142 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lsnff" event={"ID":"cbba9a43-445e-4657-9a7b-35cd2d2dc405","Type":"ContainerStarted","Data":"ac661125407b2a770c82c3a0b5090c86e4744b1cca1f278f9f93b60093b7ab28"} Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.777666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b3c8-account-create-n8x6v" event={"ID":"e25c9bc5-d0fc-48e4-9579-59b360e49731","Type":"ContainerStarted","Data":"a3199cceb1a1590bb1f9f6496de156eed1e565bfdb2b9afd81cb3775dbe9f5ec"} Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.777697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b3c8-account-create-n8x6v" event={"ID":"e25c9bc5-d0fc-48e4-9579-59b360e49731","Type":"ContainerStarted","Data":"49bb1c563d0055ada47f8c1b1a6a60a10ca8faaeb600a509998c42c81fb99b52"} Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.781087 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3410-account-create-c74k2" event={"ID":"01ac49a1-753f-46bf-bddd-410c28fe864e","Type":"ContainerStarted","Data":"fc12f9e41717cf3b6f976eb4a1b63c30c565fc37de6118c68070f24a1120cee2"} Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.781122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3410-account-create-c74k2" event={"ID":"01ac49a1-753f-46bf-bddd-410c28fe864e","Type":"ContainerStarted","Data":"65fb31e35dc837c931c122ee18e449d452d8e62743e41017b02d5e871b264616"} Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.783580 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2gs6v" event={"ID":"827adb53-2cb6-46ec-be78-6d25ced71cff","Type":"ContainerStarted","Data":"7fc1080cbb820bcb5d38751e729538916b95a8ba3d604c89c2eaf90a17f24b73"} Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.783630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2gs6v" event={"ID":"827adb53-2cb6-46ec-be78-6d25ced71cff","Type":"ContainerStarted","Data":"635fe4e241e4c807ef77213e5d3008630a78d634f04c14b9e103b6c008d39e79"} Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.793789 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-lsnff" podStartSLOduration=1.793775385 podStartE2EDuration="1.793775385s" podCreationTimestamp="2025-11-24 13:32:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:52.793308483 +0000 UTC m=+1114.432847783" watchObservedRunningTime="2025-11-24 13:32:52.793775385 +0000 UTC m=+1114.433314695" Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.827653 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-b3c8-account-create-n8x6v" podStartSLOduration=1.8276356169999999 podStartE2EDuration="1.827635617s" podCreationTimestamp="2025-11-24 13:32:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:52.811401919 +0000 UTC m=+1114.450941219" watchObservedRunningTime="2025-11-24 13:32:52.827635617 +0000 UTC m=+1114.467174927" Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.829833 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-2gs6v" podStartSLOduration=1.8298230229999999 podStartE2EDuration="1.829823023s" podCreationTimestamp="2025-11-24 13:32:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:52.824359183 +0000 UTC m=+1114.463898493" watchObservedRunningTime="2025-11-24 13:32:52.829823023 +0000 UTC m=+1114.469362333" Nov 24 13:32:52 crc kubenswrapper[4824]: I1124 13:32:52.845245 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-3410-account-create-c74k2" podStartSLOduration=1.84522838 podStartE2EDuration="1.84522838s" podCreationTimestamp="2025-11-24 13:32:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:52.841037472 +0000 UTC m=+1114.480576782" watchObservedRunningTime="2025-11-24 13:32:52.84522838 +0000 UTC m=+1114.484767690" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.095969 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.179255 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2cfsp"] Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.179532 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" podUID="f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" containerName="dnsmasq-dns" containerID="cri-o://742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb" gracePeriod=10 Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.252900 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-74w28"] Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.255281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.279161 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-74w28"] Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.375051 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-dns-svc\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.375350 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wstfq\" (UniqueName: \"kubernetes.io/projected/429c7611-6a69-43a2-8047-d46b48f48447-kube-api-access-wstfq\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.375387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.375411 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-config\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.375508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.477127 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-dns-svc\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.478261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wstfq\" (UniqueName: \"kubernetes.io/projected/429c7611-6a69-43a2-8047-d46b48f48447-kube-api-access-wstfq\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.478201 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-dns-svc\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.478343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.479185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.482492 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-config\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.483068 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-config\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.483143 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.483725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.500355 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wstfq\" (UniqueName: \"kubernetes.io/projected/429c7611-6a69-43a2-8047-d46b48f48447-kube-api-access-wstfq\") pod \"dnsmasq-dns-698758b865-74w28\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.638281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.673118 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.787222 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-config\") pod \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.787321 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlm4m\" (UniqueName: \"kubernetes.io/projected/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-kube-api-access-zlm4m\") pod \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.787371 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-ovsdbserver-nb\") pod \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.787533 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-dns-svc\") pod \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\" (UID: \"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f\") " Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.807800 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-kube-api-access-zlm4m" (OuterVolumeSpecName: "kube-api-access-zlm4m") pod "f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" (UID: "f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f"). InnerVolumeSpecName "kube-api-access-zlm4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.822452 4824 generic.go:334] "Generic (PLEG): container finished" podID="01ac49a1-753f-46bf-bddd-410c28fe864e" containerID="fc12f9e41717cf3b6f976eb4a1b63c30c565fc37de6118c68070f24a1120cee2" exitCode=0 Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.822685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3410-account-create-c74k2" event={"ID":"01ac49a1-753f-46bf-bddd-410c28fe864e","Type":"ContainerDied","Data":"fc12f9e41717cf3b6f976eb4a1b63c30c565fc37de6118c68070f24a1120cee2"} Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.833082 4824 generic.go:334] "Generic (PLEG): container finished" podID="827adb53-2cb6-46ec-be78-6d25ced71cff" containerID="7fc1080cbb820bcb5d38751e729538916b95a8ba3d604c89c2eaf90a17f24b73" exitCode=0 Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.833161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2gs6v" event={"ID":"827adb53-2cb6-46ec-be78-6d25ced71cff","Type":"ContainerDied","Data":"7fc1080cbb820bcb5d38751e729538916b95a8ba3d604c89c2eaf90a17f24b73"} Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.840143 4824 generic.go:334] "Generic (PLEG): container finished" podID="cbba9a43-445e-4657-9a7b-35cd2d2dc405" containerID="987c50051bd0e9f80f7397d64a85e667b6a2c95451fc31064e48afcda69b206b" exitCode=0 Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.840245 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lsnff" event={"ID":"cbba9a43-445e-4657-9a7b-35cd2d2dc405","Type":"ContainerDied","Data":"987c50051bd0e9f80f7397d64a85e667b6a2c95451fc31064e48afcda69b206b"} Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.845633 4824 generic.go:334] "Generic (PLEG): container finished" podID="f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" containerID="742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb" exitCode=0 Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.845698 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" event={"ID":"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f","Type":"ContainerDied","Data":"742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb"} Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.845723 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" event={"ID":"f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f","Type":"ContainerDied","Data":"4f867d156b7f9c17122e455ceea67f0cb883d092774f988e7885993834927f9e"} Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.845739 4824 scope.go:117] "RemoveContainer" containerID="742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.845850 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2cfsp" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.849406 4824 generic.go:334] "Generic (PLEG): container finished" podID="e25c9bc5-d0fc-48e4-9579-59b360e49731" containerID="a3199cceb1a1590bb1f9f6496de156eed1e565bfdb2b9afd81cb3775dbe9f5ec" exitCode=0 Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.849493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b3c8-account-create-n8x6v" event={"ID":"e25c9bc5-d0fc-48e4-9579-59b360e49731","Type":"ContainerDied","Data":"a3199cceb1a1590bb1f9f6496de156eed1e565bfdb2b9afd81cb3775dbe9f5ec"} Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.879715 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" (UID: "f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.882903 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" (UID: "f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.883122 4824 scope.go:117] "RemoveContainer" containerID="605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.889754 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlm4m\" (UniqueName: \"kubernetes.io/projected/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-kube-api-access-zlm4m\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.889774 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.889784 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.918945 4824 scope.go:117] "RemoveContainer" containerID="742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb" Nov 24 13:32:53 crc kubenswrapper[4824]: E1124 13:32:53.922104 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb\": container with ID starting with 742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb not found: ID does not exist" containerID="742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.922158 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb"} err="failed to get container status \"742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb\": rpc error: code = NotFound desc = could not find container \"742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb\": container with ID starting with 742d2cbfaa18c6f24a9a744d10c969946107cc84383737cead2010b3867681eb not found: ID does not exist" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.922183 4824 scope.go:117] "RemoveContainer" containerID="605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8" Nov 24 13:32:53 crc kubenswrapper[4824]: E1124 13:32:53.925363 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8\": container with ID starting with 605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8 not found: ID does not exist" containerID="605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.925404 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8"} err="failed to get container status \"605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8\": rpc error: code = NotFound desc = could not find container \"605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8\": container with ID starting with 605be70a7f289cd1ddca9be2311d59c0be05e411ef0fece8563397e3f58a43a8 not found: ID does not exist" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.934694 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-config" (OuterVolumeSpecName: "config") pod "f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" (UID: "f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:53 crc kubenswrapper[4824]: I1124 13:32:53.991700 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.173223 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2cfsp"] Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.178209 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2cfsp"] Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.183323 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-74w28"] Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.274628 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 13:32:54 crc kubenswrapper[4824]: E1124 13:32:54.274968 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" containerName="init" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.274987 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" containerName="init" Nov 24 13:32:54 crc kubenswrapper[4824]: E1124 13:32:54.275003 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" containerName="dnsmasq-dns" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.275011 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" containerName="dnsmasq-dns" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.275173 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" containerName="dnsmasq-dns" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.279447 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.286420 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.286697 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-xk9l8" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.286956 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.286965 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.295317 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.396941 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-lock\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.396992 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.397025 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.397065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q7mp\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-kube-api-access-8q7mp\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.397135 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-cache\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.498667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q7mp\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-kube-api-access-8q7mp\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.498782 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-cache\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.498881 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-lock\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.498915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.498954 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: E1124 13:32:54.499089 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:32:54 crc kubenswrapper[4824]: E1124 13:32:54.499134 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:32:54 crc kubenswrapper[4824]: E1124 13:32:54.499187 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift podName:1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c nodeName:}" failed. No retries permitted until 2025-11-24 13:32:54.99916894 +0000 UTC m=+1116.638708250 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift") pod "swift-storage-0" (UID: "1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c") : configmap "swift-ring-files" not found Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.499290 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-cache\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.499387 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.499401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-lock\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.517708 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q7mp\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-kube-api-access-8q7mp\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.520607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.782913 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-m5pdt"] Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.784706 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.793203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.797186 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.806149 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.842193 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m5pdt"] Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.889827 4824 generic.go:334] "Generic (PLEG): container finished" podID="429c7611-6a69-43a2-8047-d46b48f48447" containerID="27a099e0535ad549e5b39e4258eca640dc27d8c79443cd5ceccb564ff76634b1" exitCode=0 Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.891404 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-74w28" event={"ID":"429c7611-6a69-43a2-8047-d46b48f48447","Type":"ContainerDied","Data":"27a099e0535ad549e5b39e4258eca640dc27d8c79443cd5ceccb564ff76634b1"} Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.891513 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-74w28" event={"ID":"429c7611-6a69-43a2-8047-d46b48f48447","Type":"ContainerStarted","Data":"81f14dd73e3f1b0e26b8ffb4b8c9526d8f8c77b94beee9b76eda0d9f7d12b90f"} Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.906960 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-ring-data-devices\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.907042 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-scripts\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.907081 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c2abd860-636d-4e37-8a85-334aefcf44ec-etc-swift\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.907139 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-swiftconf\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.907188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq669\" (UniqueName: \"kubernetes.io/projected/c2abd860-636d-4e37-8a85-334aefcf44ec-kube-api-access-hq669\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.907219 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-dispersionconf\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:54 crc kubenswrapper[4824]: I1124 13:32:54.907238 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-combined-ca-bundle\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.008329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-swiftconf\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.008406 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq669\" (UniqueName: \"kubernetes.io/projected/c2abd860-636d-4e37-8a85-334aefcf44ec-kube-api-access-hq669\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.008439 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-dispersionconf\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.008454 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-combined-ca-bundle\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.008519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.008550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-ring-data-devices\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.008582 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-scripts\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.008605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c2abd860-636d-4e37-8a85-334aefcf44ec-etc-swift\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.009018 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c2abd860-636d-4e37-8a85-334aefcf44ec-etc-swift\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.012401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-swiftconf\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: E1124 13:32:55.012622 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:32:55 crc kubenswrapper[4824]: E1124 13:32:55.012637 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:32:55 crc kubenswrapper[4824]: E1124 13:32:55.012684 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift podName:1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c nodeName:}" failed. No retries permitted until 2025-11-24 13:32:56.012670123 +0000 UTC m=+1117.652209423 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift") pod "swift-storage-0" (UID: "1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c") : configmap "swift-ring-files" not found Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.013580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-ring-data-devices\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.013978 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-scripts\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.020273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-dispersionconf\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.031924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-combined-ca-bundle\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.033305 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq669\" (UniqueName: \"kubernetes.io/projected/c2abd860-636d-4e37-8a85-334aefcf44ec-kube-api-access-hq669\") pod \"swift-ring-rebalance-m5pdt\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.052457 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f" path="/var/lib/kubelet/pods/f94e4c2e-4fce-40b1-ab5c-c4fda5dd7a5f/volumes" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.138657 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.298141 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lsnff" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.413161 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.419916 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmq7h\" (UniqueName: \"kubernetes.io/projected/cbba9a43-445e-4657-9a7b-35cd2d2dc405-kube-api-access-dmq7h\") pod \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\" (UID: \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\") " Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.420106 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba9a43-445e-4657-9a7b-35cd2d2dc405-operator-scripts\") pod \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\" (UID: \"cbba9a43-445e-4657-9a7b-35cd2d2dc405\") " Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.421153 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbba9a43-445e-4657-9a7b-35cd2d2dc405-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cbba9a43-445e-4657-9a7b-35cd2d2dc405" (UID: "cbba9a43-445e-4657-9a7b-35cd2d2dc405"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.433339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbba9a43-445e-4657-9a7b-35cd2d2dc405-kube-api-access-dmq7h" (OuterVolumeSpecName: "kube-api-access-dmq7h") pod "cbba9a43-445e-4657-9a7b-35cd2d2dc405" (UID: "cbba9a43-445e-4657-9a7b-35cd2d2dc405"). InnerVolumeSpecName "kube-api-access-dmq7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.521471 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25c9bc5-d0fc-48e4-9579-59b360e49731-operator-scripts\") pod \"e25c9bc5-d0fc-48e4-9579-59b360e49731\" (UID: \"e25c9bc5-d0fc-48e4-9579-59b360e49731\") " Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.521798 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb47c\" (UniqueName: \"kubernetes.io/projected/e25c9bc5-d0fc-48e4-9579-59b360e49731-kube-api-access-wb47c\") pod \"e25c9bc5-d0fc-48e4-9579-59b360e49731\" (UID: \"e25c9bc5-d0fc-48e4-9579-59b360e49731\") " Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.522330 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e25c9bc5-d0fc-48e4-9579-59b360e49731-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e25c9bc5-d0fc-48e4-9579-59b360e49731" (UID: "e25c9bc5-d0fc-48e4-9579-59b360e49731"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.523197 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmq7h\" (UniqueName: \"kubernetes.io/projected/cbba9a43-445e-4657-9a7b-35cd2d2dc405-kube-api-access-dmq7h\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.523228 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25c9bc5-d0fc-48e4-9579-59b360e49731-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.523237 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba9a43-445e-4657-9a7b-35cd2d2dc405-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.526285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e25c9bc5-d0fc-48e4-9579-59b360e49731-kube-api-access-wb47c" (OuterVolumeSpecName: "kube-api-access-wb47c") pod "e25c9bc5-d0fc-48e4-9579-59b360e49731" (UID: "e25c9bc5-d0fc-48e4-9579-59b360e49731"). InnerVolumeSpecName "kube-api-access-wb47c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.590720 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.624471 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb47c\" (UniqueName: \"kubernetes.io/projected/e25c9bc5-d0fc-48e4-9579-59b360e49731-kube-api-access-wb47c\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.635862 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.726129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp6jb\" (UniqueName: \"kubernetes.io/projected/01ac49a1-753f-46bf-bddd-410c28fe864e-kube-api-access-dp6jb\") pod \"01ac49a1-753f-46bf-bddd-410c28fe864e\" (UID: \"01ac49a1-753f-46bf-bddd-410c28fe864e\") " Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.726229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/827adb53-2cb6-46ec-be78-6d25ced71cff-operator-scripts\") pod \"827adb53-2cb6-46ec-be78-6d25ced71cff\" (UID: \"827adb53-2cb6-46ec-be78-6d25ced71cff\") " Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.726289 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ac49a1-753f-46bf-bddd-410c28fe864e-operator-scripts\") pod \"01ac49a1-753f-46bf-bddd-410c28fe864e\" (UID: \"01ac49a1-753f-46bf-bddd-410c28fe864e\") " Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.726346 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rclz\" (UniqueName: \"kubernetes.io/projected/827adb53-2cb6-46ec-be78-6d25ced71cff-kube-api-access-8rclz\") pod \"827adb53-2cb6-46ec-be78-6d25ced71cff\" (UID: \"827adb53-2cb6-46ec-be78-6d25ced71cff\") " Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.727413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/827adb53-2cb6-46ec-be78-6d25ced71cff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "827adb53-2cb6-46ec-be78-6d25ced71cff" (UID: "827adb53-2cb6-46ec-be78-6d25ced71cff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.727562 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ac49a1-753f-46bf-bddd-410c28fe864e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "01ac49a1-753f-46bf-bddd-410c28fe864e" (UID: "01ac49a1-753f-46bf-bddd-410c28fe864e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.730850 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ac49a1-753f-46bf-bddd-410c28fe864e-kube-api-access-dp6jb" (OuterVolumeSpecName: "kube-api-access-dp6jb") pod "01ac49a1-753f-46bf-bddd-410c28fe864e" (UID: "01ac49a1-753f-46bf-bddd-410c28fe864e"). InnerVolumeSpecName "kube-api-access-dp6jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.730957 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/827adb53-2cb6-46ec-be78-6d25ced71cff-kube-api-access-8rclz" (OuterVolumeSpecName: "kube-api-access-8rclz") pod "827adb53-2cb6-46ec-be78-6d25ced71cff" (UID: "827adb53-2cb6-46ec-be78-6d25ced71cff"). InnerVolumeSpecName "kube-api-access-8rclz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.770627 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m5pdt"] Nov 24 13:32:55 crc kubenswrapper[4824]: W1124 13:32:55.777974 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2abd860_636d_4e37_8a85_334aefcf44ec.slice/crio-a9020a37bd27ecfcc9c8eb97b37b57b2a6c43b05726062c6b621a51c76d34dfb WatchSource:0}: Error finding container a9020a37bd27ecfcc9c8eb97b37b57b2a6c43b05726062c6b621a51c76d34dfb: Status 404 returned error can't find the container with id a9020a37bd27ecfcc9c8eb97b37b57b2a6c43b05726062c6b621a51c76d34dfb Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.828309 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/827adb53-2cb6-46ec-be78-6d25ced71cff-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.828345 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01ac49a1-753f-46bf-bddd-410c28fe864e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.828355 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rclz\" (UniqueName: \"kubernetes.io/projected/827adb53-2cb6-46ec-be78-6d25ced71cff-kube-api-access-8rclz\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.828364 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp6jb\" (UniqueName: \"kubernetes.io/projected/01ac49a1-753f-46bf-bddd-410c28fe864e-kube-api-access-dp6jb\") on node \"crc\" DevicePath \"\"" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.897118 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m5pdt" event={"ID":"c2abd860-636d-4e37-8a85-334aefcf44ec","Type":"ContainerStarted","Data":"a9020a37bd27ecfcc9c8eb97b37b57b2a6c43b05726062c6b621a51c76d34dfb"} Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.898733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b3c8-account-create-n8x6v" event={"ID":"e25c9bc5-d0fc-48e4-9579-59b360e49731","Type":"ContainerDied","Data":"49bb1c563d0055ada47f8c1b1a6a60a10ca8faaeb600a509998c42c81fb99b52"} Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.898758 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49bb1c563d0055ada47f8c1b1a6a60a10ca8faaeb600a509998c42c81fb99b52" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.898791 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b3c8-account-create-n8x6v" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.900050 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-74w28" event={"ID":"429c7611-6a69-43a2-8047-d46b48f48447","Type":"ContainerStarted","Data":"0eacd6c6d75a66c6c694c6209d042d8d8cefc528e587d8758df40d15b776d18d"} Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.900903 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.902113 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3410-account-create-c74k2" event={"ID":"01ac49a1-753f-46bf-bddd-410c28fe864e","Type":"ContainerDied","Data":"65fb31e35dc837c931c122ee18e449d452d8e62743e41017b02d5e871b264616"} Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.902133 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65fb31e35dc837c931c122ee18e449d452d8e62743e41017b02d5e871b264616" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.902165 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3410-account-create-c74k2" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.909826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2gs6v" event={"ID":"827adb53-2cb6-46ec-be78-6d25ced71cff","Type":"ContainerDied","Data":"635fe4e241e4c807ef77213e5d3008630a78d634f04c14b9e103b6c008d39e79"} Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.909857 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="635fe4e241e4c807ef77213e5d3008630a78d634f04c14b9e103b6c008d39e79" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.909862 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2gs6v" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.911425 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lsnff" event={"ID":"cbba9a43-445e-4657-9a7b-35cd2d2dc405","Type":"ContainerDied","Data":"ac661125407b2a770c82c3a0b5090c86e4744b1cca1f278f9f93b60093b7ab28"} Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.911442 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac661125407b2a770c82c3a0b5090c86e4744b1cca1f278f9f93b60093b7ab28" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.911496 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lsnff" Nov 24 13:32:55 crc kubenswrapper[4824]: I1124 13:32:55.934260 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-74w28" podStartSLOduration=2.934240473 podStartE2EDuration="2.934240473s" podCreationTimestamp="2025-11-24 13:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:32:55.928313341 +0000 UTC m=+1117.567852661" watchObservedRunningTime="2025-11-24 13:32:55.934240473 +0000 UTC m=+1117.573779783" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.031135 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:56 crc kubenswrapper[4824]: E1124 13:32:56.031390 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:32:56 crc kubenswrapper[4824]: E1124 13:32:56.031425 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:32:56 crc kubenswrapper[4824]: E1124 13:32:56.031483 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift podName:1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c nodeName:}" failed. No retries permitted until 2025-11-24 13:32:58.031462437 +0000 UTC m=+1119.671001757 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift") pod "swift-storage-0" (UID: "1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c") : configmap "swift-ring-files" not found Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.882883 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-scfqc"] Nov 24 13:32:56 crc kubenswrapper[4824]: E1124 13:32:56.883215 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ac49a1-753f-46bf-bddd-410c28fe864e" containerName="mariadb-account-create" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.883229 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ac49a1-753f-46bf-bddd-410c28fe864e" containerName="mariadb-account-create" Nov 24 13:32:56 crc kubenswrapper[4824]: E1124 13:32:56.883242 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="827adb53-2cb6-46ec-be78-6d25ced71cff" containerName="mariadb-database-create" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.883249 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="827adb53-2cb6-46ec-be78-6d25ced71cff" containerName="mariadb-database-create" Nov 24 13:32:56 crc kubenswrapper[4824]: E1124 13:32:56.883260 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e25c9bc5-d0fc-48e4-9579-59b360e49731" containerName="mariadb-account-create" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.883267 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e25c9bc5-d0fc-48e4-9579-59b360e49731" containerName="mariadb-account-create" Nov 24 13:32:56 crc kubenswrapper[4824]: E1124 13:32:56.883275 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbba9a43-445e-4657-9a7b-35cd2d2dc405" containerName="mariadb-database-create" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.883281 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbba9a43-445e-4657-9a7b-35cd2d2dc405" containerName="mariadb-database-create" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.883447 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e25c9bc5-d0fc-48e4-9579-59b360e49731" containerName="mariadb-account-create" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.883457 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="827adb53-2cb6-46ec-be78-6d25ced71cff" containerName="mariadb-database-create" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.883470 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ac49a1-753f-46bf-bddd-410c28fe864e" containerName="mariadb-account-create" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.883484 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbba9a43-445e-4657-9a7b-35cd2d2dc405" containerName="mariadb-database-create" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.883974 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-scfqc" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.895282 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-scfqc"] Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.947402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l8zd\" (UniqueName: \"kubernetes.io/projected/5ac2f711-f52c-4424-930b-a40967d136c6-kube-api-access-8l8zd\") pod \"glance-db-create-scfqc\" (UID: \"5ac2f711-f52c-4424-930b-a40967d136c6\") " pod="openstack/glance-db-create-scfqc" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.947490 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ac2f711-f52c-4424-930b-a40967d136c6-operator-scripts\") pod \"glance-db-create-scfqc\" (UID: \"5ac2f711-f52c-4424-930b-a40967d136c6\") " pod="openstack/glance-db-create-scfqc" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.990955 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6389-account-create-2pp9j"] Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.992300 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:32:56 crc kubenswrapper[4824]: I1124 13:32:56.993984 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.034871 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6389-account-create-2pp9j"] Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.048826 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l8zd\" (UniqueName: \"kubernetes.io/projected/5ac2f711-f52c-4424-930b-a40967d136c6-kube-api-access-8l8zd\") pod \"glance-db-create-scfqc\" (UID: \"5ac2f711-f52c-4424-930b-a40967d136c6\") " pod="openstack/glance-db-create-scfqc" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.048941 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6faec986-3929-49c0-98e4-021352f2a8b9-operator-scripts\") pod \"glance-6389-account-create-2pp9j\" (UID: \"6faec986-3929-49c0-98e4-021352f2a8b9\") " pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.048959 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p58wt\" (UniqueName: \"kubernetes.io/projected/6faec986-3929-49c0-98e4-021352f2a8b9-kube-api-access-p58wt\") pod \"glance-6389-account-create-2pp9j\" (UID: \"6faec986-3929-49c0-98e4-021352f2a8b9\") " pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.049239 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ac2f711-f52c-4424-930b-a40967d136c6-operator-scripts\") pod \"glance-db-create-scfqc\" (UID: \"5ac2f711-f52c-4424-930b-a40967d136c6\") " pod="openstack/glance-db-create-scfqc" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.049876 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ac2f711-f52c-4424-930b-a40967d136c6-operator-scripts\") pod \"glance-db-create-scfqc\" (UID: \"5ac2f711-f52c-4424-930b-a40967d136c6\") " pod="openstack/glance-db-create-scfqc" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.080921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l8zd\" (UniqueName: \"kubernetes.io/projected/5ac2f711-f52c-4424-930b-a40967d136c6-kube-api-access-8l8zd\") pod \"glance-db-create-scfqc\" (UID: \"5ac2f711-f52c-4424-930b-a40967d136c6\") " pod="openstack/glance-db-create-scfqc" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.153017 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6faec986-3929-49c0-98e4-021352f2a8b9-operator-scripts\") pod \"glance-6389-account-create-2pp9j\" (UID: \"6faec986-3929-49c0-98e4-021352f2a8b9\") " pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.153067 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p58wt\" (UniqueName: \"kubernetes.io/projected/6faec986-3929-49c0-98e4-021352f2a8b9-kube-api-access-p58wt\") pod \"glance-6389-account-create-2pp9j\" (UID: \"6faec986-3929-49c0-98e4-021352f2a8b9\") " pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.154519 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6faec986-3929-49c0-98e4-021352f2a8b9-operator-scripts\") pod \"glance-6389-account-create-2pp9j\" (UID: \"6faec986-3929-49c0-98e4-021352f2a8b9\") " pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.187470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p58wt\" (UniqueName: \"kubernetes.io/projected/6faec986-3929-49c0-98e4-021352f2a8b9-kube-api-access-p58wt\") pod \"glance-6389-account-create-2pp9j\" (UID: \"6faec986-3929-49c0-98e4-021352f2a8b9\") " pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.207891 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-scfqc" Nov 24 13:32:57 crc kubenswrapper[4824]: I1124 13:32:57.310147 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:32:58 crc kubenswrapper[4824]: I1124 13:32:58.071070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:32:58 crc kubenswrapper[4824]: E1124 13:32:58.071942 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:32:58 crc kubenswrapper[4824]: E1124 13:32:58.071968 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:32:58 crc kubenswrapper[4824]: E1124 13:32:58.072008 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift podName:1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c nodeName:}" failed. No retries permitted until 2025-11-24 13:33:02.071993472 +0000 UTC m=+1123.711532782 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift") pod "swift-storage-0" (UID: "1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c") : configmap "swift-ring-files" not found Nov 24 13:32:58 crc kubenswrapper[4824]: I1124 13:32:58.160989 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:32:59 crc kubenswrapper[4824]: I1124 13:32:59.871750 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-scfqc"] Nov 24 13:32:59 crc kubenswrapper[4824]: W1124 13:32:59.895558 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6faec986_3929_49c0_98e4_021352f2a8b9.slice/crio-844065f6ea990907bd6944ab7289d2c6aa0df9e2ca8c7a57a1d0e8a4707c86ed WatchSource:0}: Error finding container 844065f6ea990907bd6944ab7289d2c6aa0df9e2ca8c7a57a1d0e8a4707c86ed: Status 404 returned error can't find the container with id 844065f6ea990907bd6944ab7289d2c6aa0df9e2ca8c7a57a1d0e8a4707c86ed Nov 24 13:32:59 crc kubenswrapper[4824]: I1124 13:32:59.896845 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6389-account-create-2pp9j"] Nov 24 13:32:59 crc kubenswrapper[4824]: I1124 13:32:59.951900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6389-account-create-2pp9j" event={"ID":"6faec986-3929-49c0-98e4-021352f2a8b9","Type":"ContainerStarted","Data":"844065f6ea990907bd6944ab7289d2c6aa0df9e2ca8c7a57a1d0e8a4707c86ed"} Nov 24 13:32:59 crc kubenswrapper[4824]: I1124 13:32:59.959353 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m5pdt" event={"ID":"c2abd860-636d-4e37-8a85-334aefcf44ec","Type":"ContainerStarted","Data":"0635fcf7c3af6a50234fd8c12832747e790d4d1d0c814a1c185a4582ed031ffc"} Nov 24 13:32:59 crc kubenswrapper[4824]: I1124 13:32:59.963000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-scfqc" event={"ID":"5ac2f711-f52c-4424-930b-a40967d136c6","Type":"ContainerStarted","Data":"7a8ebc6509d8ff3599c1092a8de6007f1682e3dcb0c7859940f1be8e3d8f3e56"} Nov 24 13:32:59 crc kubenswrapper[4824]: I1124 13:32:59.978194 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-m5pdt" podStartSLOduration=2.359923995 podStartE2EDuration="5.978176587s" podCreationTimestamp="2025-11-24 13:32:54 +0000 UTC" firstStartedPulling="2025-11-24 13:32:55.780546106 +0000 UTC m=+1117.420085436" lastFinishedPulling="2025-11-24 13:32:59.398798708 +0000 UTC m=+1121.038338028" observedRunningTime="2025-11-24 13:32:59.977626603 +0000 UTC m=+1121.617165913" watchObservedRunningTime="2025-11-24 13:32:59.978176587 +0000 UTC m=+1121.617715897" Nov 24 13:33:00 crc kubenswrapper[4824]: I1124 13:33:00.976979 4824 generic.go:334] "Generic (PLEG): container finished" podID="5ac2f711-f52c-4424-930b-a40967d136c6" containerID="756cabcd5fa409c5ef99a5b9bd605d437bf5427e3e473f4d97178ff72772029a" exitCode=0 Nov 24 13:33:00 crc kubenswrapper[4824]: I1124 13:33:00.977032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-scfqc" event={"ID":"5ac2f711-f52c-4424-930b-a40967d136c6","Type":"ContainerDied","Data":"756cabcd5fa409c5ef99a5b9bd605d437bf5427e3e473f4d97178ff72772029a"} Nov 24 13:33:00 crc kubenswrapper[4824]: I1124 13:33:00.981253 4824 generic.go:334] "Generic (PLEG): container finished" podID="6faec986-3929-49c0-98e4-021352f2a8b9" containerID="14383d0ba2a9e4d43349c9d7a0c19a2023268e801a841a423c5134f7f69ed804" exitCode=0 Nov 24 13:33:00 crc kubenswrapper[4824]: I1124 13:33:00.981330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6389-account-create-2pp9j" event={"ID":"6faec986-3929-49c0-98e4-021352f2a8b9","Type":"ContainerDied","Data":"14383d0ba2a9e4d43349c9d7a0c19a2023268e801a841a423c5134f7f69ed804"} Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.154265 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:33:02 crc kubenswrapper[4824]: E1124 13:33:02.154918 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 13:33:02 crc kubenswrapper[4824]: E1124 13:33:02.154957 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 13:33:02 crc kubenswrapper[4824]: E1124 13:33:02.155030 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift podName:1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c nodeName:}" failed. No retries permitted until 2025-11-24 13:33:10.155008512 +0000 UTC m=+1131.794547832 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift") pod "swift-storage-0" (UID: "1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c") : configmap "swift-ring-files" not found Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.393019 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.393400 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-scfqc" Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.458791 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6faec986-3929-49c0-98e4-021352f2a8b9-operator-scripts\") pod \"6faec986-3929-49c0-98e4-021352f2a8b9\" (UID: \"6faec986-3929-49c0-98e4-021352f2a8b9\") " Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.458950 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ac2f711-f52c-4424-930b-a40967d136c6-operator-scripts\") pod \"5ac2f711-f52c-4424-930b-a40967d136c6\" (UID: \"5ac2f711-f52c-4424-930b-a40967d136c6\") " Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.459003 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l8zd\" (UniqueName: \"kubernetes.io/projected/5ac2f711-f52c-4424-930b-a40967d136c6-kube-api-access-8l8zd\") pod \"5ac2f711-f52c-4424-930b-a40967d136c6\" (UID: \"5ac2f711-f52c-4424-930b-a40967d136c6\") " Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.459081 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p58wt\" (UniqueName: \"kubernetes.io/projected/6faec986-3929-49c0-98e4-021352f2a8b9-kube-api-access-p58wt\") pod \"6faec986-3929-49c0-98e4-021352f2a8b9\" (UID: \"6faec986-3929-49c0-98e4-021352f2a8b9\") " Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.459577 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ac2f711-f52c-4424-930b-a40967d136c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ac2f711-f52c-4424-930b-a40967d136c6" (UID: "5ac2f711-f52c-4424-930b-a40967d136c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.460070 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6faec986-3929-49c0-98e4-021352f2a8b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6faec986-3929-49c0-98e4-021352f2a8b9" (UID: "6faec986-3929-49c0-98e4-021352f2a8b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.466154 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6faec986-3929-49c0-98e4-021352f2a8b9-kube-api-access-p58wt" (OuterVolumeSpecName: "kube-api-access-p58wt") pod "6faec986-3929-49c0-98e4-021352f2a8b9" (UID: "6faec986-3929-49c0-98e4-021352f2a8b9"). InnerVolumeSpecName "kube-api-access-p58wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.466657 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ac2f711-f52c-4424-930b-a40967d136c6-kube-api-access-8l8zd" (OuterVolumeSpecName: "kube-api-access-8l8zd") pod "5ac2f711-f52c-4424-930b-a40967d136c6" (UID: "5ac2f711-f52c-4424-930b-a40967d136c6"). InnerVolumeSpecName "kube-api-access-8l8zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.561048 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p58wt\" (UniqueName: \"kubernetes.io/projected/6faec986-3929-49c0-98e4-021352f2a8b9-kube-api-access-p58wt\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.561110 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6faec986-3929-49c0-98e4-021352f2a8b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.561119 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ac2f711-f52c-4424-930b-a40967d136c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:02 crc kubenswrapper[4824]: I1124 13:33:02.561128 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l8zd\" (UniqueName: \"kubernetes.io/projected/5ac2f711-f52c-4424-930b-a40967d136c6-kube-api-access-8l8zd\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.007068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-scfqc" event={"ID":"5ac2f711-f52c-4424-930b-a40967d136c6","Type":"ContainerDied","Data":"7a8ebc6509d8ff3599c1092a8de6007f1682e3dcb0c7859940f1be8e3d8f3e56"} Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.007096 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-scfqc" Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.007113 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a8ebc6509d8ff3599c1092a8de6007f1682e3dcb0c7859940f1be8e3d8f3e56" Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.018524 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6389-account-create-2pp9j" Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.035638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6389-account-create-2pp9j" event={"ID":"6faec986-3929-49c0-98e4-021352f2a8b9","Type":"ContainerDied","Data":"844065f6ea990907bd6944ab7289d2c6aa0df9e2ca8c7a57a1d0e8a4707c86ed"} Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.035693 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="844065f6ea990907bd6944ab7289d2c6aa0df9e2ca8c7a57a1d0e8a4707c86ed" Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.521519 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.640058 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.700886 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qdtw7"] Nov 24 13:33:03 crc kubenswrapper[4824]: I1124 13:33:03.706245 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" podUID="e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" containerName="dnsmasq-dns" containerID="cri-o://201ed3365f0b89394e4fa4fb5ecf036ea8cd108265816cac747fbd7a17870f85" gracePeriod=10 Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.033020 4824 generic.go:334] "Generic (PLEG): container finished" podID="e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" containerID="201ed3365f0b89394e4fa4fb5ecf036ea8cd108265816cac747fbd7a17870f85" exitCode=0 Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.033120 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" event={"ID":"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7","Type":"ContainerDied","Data":"201ed3365f0b89394e4fa4fb5ecf036ea8cd108265816cac747fbd7a17870f85"} Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.037342 4824 generic.go:334] "Generic (PLEG): container finished" podID="bc596e02-b760-49d3-a0c2-00ec8e751e05" containerID="46d8ac6888d06952ac6295da8bf804f80e93e9aa76eb2e3004f7277c99076e12" exitCode=0 Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.037440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bc596e02-b760-49d3-a0c2-00ec8e751e05","Type":"ContainerDied","Data":"46d8ac6888d06952ac6295da8bf804f80e93e9aa76eb2e3004f7277c99076e12"} Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.066397 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" containerID="3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c" exitCode=0 Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.066460 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7","Type":"ContainerDied","Data":"3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c"} Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.249090 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.397764 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nbqm\" (UniqueName: \"kubernetes.io/projected/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-kube-api-access-2nbqm\") pod \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.397829 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-config\") pod \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.397929 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-sb\") pod \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.397954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-nb\") pod \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.397977 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-dns-svc\") pod \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\" (UID: \"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7\") " Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.406046 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-kube-api-access-2nbqm" (OuterVolumeSpecName: "kube-api-access-2nbqm") pod "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" (UID: "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7"). InnerVolumeSpecName "kube-api-access-2nbqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.446470 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" (UID: "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.456095 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-config" (OuterVolumeSpecName: "config") pod "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" (UID: "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.473208 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" (UID: "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.479621 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" (UID: "e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.499615 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.499643 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.499652 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.499662 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nbqm\" (UniqueName: \"kubernetes.io/projected/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-kube-api-access-2nbqm\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:04 crc kubenswrapper[4824]: I1124 13:33:04.499672 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.075583 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bc596e02-b760-49d3-a0c2-00ec8e751e05","Type":"ContainerStarted","Data":"f95a2725a5975c99b2ebd62bef582367f706a83ba95e987fe32fd34e07d0c268"} Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.075777 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.077839 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7","Type":"ContainerStarted","Data":"c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126"} Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.078045 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.079834 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" event={"ID":"e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7","Type":"ContainerDied","Data":"e9c26b96f5c96bad83f9a2d310e4df743e1fe198f9df0dd76f7f81257199acae"} Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.079882 4824 scope.go:117] "RemoveContainer" containerID="201ed3365f0b89394e4fa4fb5ecf036ea8cd108265816cac747fbd7a17870f85" Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.079888 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qdtw7" Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.099575 4824 scope.go:117] "RemoveContainer" containerID="50f158d4f79ab9e8aeca8137010d3e8b30fc643cb0ac76010000a7c376828f75" Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.121069 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.715387877 podStartE2EDuration="1m8.121051259s" podCreationTimestamp="2025-11-24 13:31:57 +0000 UTC" firstStartedPulling="2025-11-24 13:31:59.370626522 +0000 UTC m=+1061.010165832" lastFinishedPulling="2025-11-24 13:32:30.776289864 +0000 UTC m=+1092.415829214" observedRunningTime="2025-11-24 13:33:05.118865733 +0000 UTC m=+1126.758405063" watchObservedRunningTime="2025-11-24 13:33:05.121051259 +0000 UTC m=+1126.760590569" Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.153162 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qdtw7"] Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.163270 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qdtw7"] Nov 24 13:33:05 crc kubenswrapper[4824]: I1124 13:33:05.179978 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.345610045 podStartE2EDuration="1m9.179956106s" podCreationTimestamp="2025-11-24 13:31:56 +0000 UTC" firstStartedPulling="2025-11-24 13:31:58.625689779 +0000 UTC m=+1060.265229089" lastFinishedPulling="2025-11-24 13:32:30.46003584 +0000 UTC m=+1092.099575150" observedRunningTime="2025-11-24 13:33:05.178272873 +0000 UTC m=+1126.817812183" watchObservedRunningTime="2025-11-24 13:33:05.179956106 +0000 UTC m=+1126.819495416" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.024577 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" path="/var/lib/kubelet/pods/e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7/volumes" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.098098 4824 generic.go:334] "Generic (PLEG): container finished" podID="c2abd860-636d-4e37-8a85-334aefcf44ec" containerID="0635fcf7c3af6a50234fd8c12832747e790d4d1d0c814a1c185a4582ed031ffc" exitCode=0 Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.098165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m5pdt" event={"ID":"c2abd860-636d-4e37-8a85-334aefcf44ec","Type":"ContainerDied","Data":"0635fcf7c3af6a50234fd8c12832747e790d4d1d0c814a1c185a4582ed031ffc"} Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.194310 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-sb5fc"] Nov 24 13:33:07 crc kubenswrapper[4824]: E1124 13:33:07.195041 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" containerName="dnsmasq-dns" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.195066 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" containerName="dnsmasq-dns" Nov 24 13:33:07 crc kubenswrapper[4824]: E1124 13:33:07.195083 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ac2f711-f52c-4424-930b-a40967d136c6" containerName="mariadb-database-create" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.195091 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ac2f711-f52c-4424-930b-a40967d136c6" containerName="mariadb-database-create" Nov 24 13:33:07 crc kubenswrapper[4824]: E1124 13:33:07.195104 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" containerName="init" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.195112 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" containerName="init" Nov 24 13:33:07 crc kubenswrapper[4824]: E1124 13:33:07.195133 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6faec986-3929-49c0-98e4-021352f2a8b9" containerName="mariadb-account-create" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.195141 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6faec986-3929-49c0-98e4-021352f2a8b9" containerName="mariadb-account-create" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.195340 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ac2f711-f52c-4424-930b-a40967d136c6" containerName="mariadb-database-create" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.195364 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6faec986-3929-49c0-98e4-021352f2a8b9" containerName="mariadb-account-create" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.195376 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8295b1a-dd8c-47f6-8fd3-5d7c36cfbfd7" containerName="dnsmasq-dns" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.197109 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.202223 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sb5fc"] Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.202360 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.202550 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-p7tpr" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.240533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4676c\" (UniqueName: \"kubernetes.io/projected/0b623f0c-385a-465d-9132-9406dfdec4ec-kube-api-access-4676c\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.240574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-config-data\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.240596 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-db-sync-config-data\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.240682 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-combined-ca-bundle\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.341798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-combined-ca-bundle\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.341902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4676c\" (UniqueName: \"kubernetes.io/projected/0b623f0c-385a-465d-9132-9406dfdec4ec-kube-api-access-4676c\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.341924 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-config-data\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.341944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-db-sync-config-data\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.349864 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-combined-ca-bundle\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.351253 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-db-sync-config-data\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.354285 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-config-data\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.359502 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4676c\" (UniqueName: \"kubernetes.io/projected/0b623f0c-385a-465d-9132-9406dfdec4ec-kube-api-access-4676c\") pod \"glance-db-sync-sb5fc\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:07 crc kubenswrapper[4824]: I1124 13:33:07.533996 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.063743 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sb5fc"] Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.106993 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sb5fc" event={"ID":"0b623f0c-385a-465d-9132-9406dfdec4ec","Type":"ContainerStarted","Data":"a619220c703562e02dea0a0b3e6e158f17118d6c6755fb021dedbf9478ab3363"} Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.424196 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.455765 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c2abd860-636d-4e37-8a85-334aefcf44ec-etc-swift\") pod \"c2abd860-636d-4e37-8a85-334aefcf44ec\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.455835 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-dispersionconf\") pod \"c2abd860-636d-4e37-8a85-334aefcf44ec\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.455873 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-ring-data-devices\") pod \"c2abd860-636d-4e37-8a85-334aefcf44ec\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.455903 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-combined-ca-bundle\") pod \"c2abd860-636d-4e37-8a85-334aefcf44ec\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.455953 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-swiftconf\") pod \"c2abd860-636d-4e37-8a85-334aefcf44ec\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.456055 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq669\" (UniqueName: \"kubernetes.io/projected/c2abd860-636d-4e37-8a85-334aefcf44ec-kube-api-access-hq669\") pod \"c2abd860-636d-4e37-8a85-334aefcf44ec\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.456093 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-scripts\") pod \"c2abd860-636d-4e37-8a85-334aefcf44ec\" (UID: \"c2abd860-636d-4e37-8a85-334aefcf44ec\") " Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.456425 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c2abd860-636d-4e37-8a85-334aefcf44ec" (UID: "c2abd860-636d-4e37-8a85-334aefcf44ec"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.457516 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2abd860-636d-4e37-8a85-334aefcf44ec-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c2abd860-636d-4e37-8a85-334aefcf44ec" (UID: "c2abd860-636d-4e37-8a85-334aefcf44ec"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.471326 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2abd860-636d-4e37-8a85-334aefcf44ec-kube-api-access-hq669" (OuterVolumeSpecName: "kube-api-access-hq669") pod "c2abd860-636d-4e37-8a85-334aefcf44ec" (UID: "c2abd860-636d-4e37-8a85-334aefcf44ec"). InnerVolumeSpecName "kube-api-access-hq669". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.488409 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c2abd860-636d-4e37-8a85-334aefcf44ec" (UID: "c2abd860-636d-4e37-8a85-334aefcf44ec"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.502743 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c2abd860-636d-4e37-8a85-334aefcf44ec" (UID: "c2abd860-636d-4e37-8a85-334aefcf44ec"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.531051 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2abd860-636d-4e37-8a85-334aefcf44ec" (UID: "c2abd860-636d-4e37-8a85-334aefcf44ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.543660 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-scripts" (OuterVolumeSpecName: "scripts") pod "c2abd860-636d-4e37-8a85-334aefcf44ec" (UID: "c2abd860-636d-4e37-8a85-334aefcf44ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.558283 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq669\" (UniqueName: \"kubernetes.io/projected/c2abd860-636d-4e37-8a85-334aefcf44ec-kube-api-access-hq669\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.558312 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.558320 4824 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c2abd860-636d-4e37-8a85-334aefcf44ec-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.558332 4824 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.558340 4824 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c2abd860-636d-4e37-8a85-334aefcf44ec-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.558351 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:08 crc kubenswrapper[4824]: I1124 13:33:08.558363 4824 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c2abd860-636d-4e37-8a85-334aefcf44ec-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:09 crc kubenswrapper[4824]: I1124 13:33:09.115946 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m5pdt" event={"ID":"c2abd860-636d-4e37-8a85-334aefcf44ec","Type":"ContainerDied","Data":"a9020a37bd27ecfcc9c8eb97b37b57b2a6c43b05726062c6b621a51c76d34dfb"} Nov 24 13:33:09 crc kubenswrapper[4824]: I1124 13:33:09.115986 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9020a37bd27ecfcc9c8eb97b37b57b2a6c43b05726062c6b621a51c76d34dfb" Nov 24 13:33:09 crc kubenswrapper[4824]: I1124 13:33:09.115999 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m5pdt" Nov 24 13:33:10 crc kubenswrapper[4824]: I1124 13:33:10.182529 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:33:10 crc kubenswrapper[4824]: I1124 13:33:10.203411 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c-etc-swift\") pod \"swift-storage-0\" (UID: \"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c\") " pod="openstack/swift-storage-0" Nov 24 13:33:10 crc kubenswrapper[4824]: I1124 13:33:10.250427 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 13:33:11 crc kubenswrapper[4824]: I1124 13:33:11.655177 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 13:33:11 crc kubenswrapper[4824]: W1124 13:33:11.662389 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ac7bb70_1d1a_4ef5_b5f2_134792dcec5c.slice/crio-9d6710934a9f1ee2609aff5561f66222aa8c17bfa316cee85be7a906a71c623b WatchSource:0}: Error finding container 9d6710934a9f1ee2609aff5561f66222aa8c17bfa316cee85be7a906a71c623b: Status 404 returned error can't find the container with id 9d6710934a9f1ee2609aff5561f66222aa8c17bfa316cee85be7a906a71c623b Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.143722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"9d6710934a9f1ee2609aff5561f66222aa8c17bfa316cee85be7a906a71c623b"} Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.186419 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ds8fs" podUID="b3c188e2-9a89-47fe-8efe-1187a3c81552" containerName="ovn-controller" probeResult="failure" output=< Nov 24 13:33:12 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 13:33:12 crc kubenswrapper[4824]: > Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.265625 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.265669 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jvdgl" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.475792 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ds8fs-config-s9ngk"] Nov 24 13:33:12 crc kubenswrapper[4824]: E1124 13:33:12.476138 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2abd860-636d-4e37-8a85-334aefcf44ec" containerName="swift-ring-rebalance" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.476157 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2abd860-636d-4e37-8a85-334aefcf44ec" containerName="swift-ring-rebalance" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.476314 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2abd860-636d-4e37-8a85-334aefcf44ec" containerName="swift-ring-rebalance" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.476857 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.491117 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.509013 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ds8fs-config-s9ngk"] Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.625473 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-log-ovn\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.625872 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-additional-scripts\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.625907 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run-ovn\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.625949 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-scripts\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.625974 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.626062 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmkls\" (UniqueName: \"kubernetes.io/projected/5d043533-d754-4424-937a-958f424b54dd-kube-api-access-tmkls\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.728892 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-log-ovn\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.728946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-additional-scripts\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.728965 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run-ovn\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.728993 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.729009 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-scripts\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.729066 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmkls\" (UniqueName: \"kubernetes.io/projected/5d043533-d754-4424-937a-958f424b54dd-kube-api-access-tmkls\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.729339 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-log-ovn\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.729428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.729464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run-ovn\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.730018 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-additional-scripts\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.731508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-scripts\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.760161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmkls\" (UniqueName: \"kubernetes.io/projected/5d043533-d754-4424-937a-958f424b54dd-kube-api-access-tmkls\") pod \"ovn-controller-ds8fs-config-s9ngk\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:12 crc kubenswrapper[4824]: I1124 13:33:12.796394 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:17 crc kubenswrapper[4824]: I1124 13:33:17.206615 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ds8fs" podUID="b3c188e2-9a89-47fe-8efe-1187a3c81552" containerName="ovn-controller" probeResult="failure" output=< Nov 24 13:33:17 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 13:33:17 crc kubenswrapper[4824]: > Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.111052 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.406443 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-ltp6q"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.407414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.420478 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ltp6q"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.436633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v822r\" (UniqueName: \"kubernetes.io/projected/77e59dd3-1218-4204-8573-24c3fa35d41b-kube-api-access-v822r\") pod \"cinder-db-create-ltp6q\" (UID: \"77e59dd3-1218-4204-8573-24c3fa35d41b\") " pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.436708 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e59dd3-1218-4204-8573-24c3fa35d41b-operator-scripts\") pod \"cinder-db-create-ltp6q\" (UID: \"77e59dd3-1218-4204-8573-24c3fa35d41b\") " pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.536048 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-nzwdl"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.537176 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.537555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v822r\" (UniqueName: \"kubernetes.io/projected/77e59dd3-1218-4204-8573-24c3fa35d41b-kube-api-access-v822r\") pod \"cinder-db-create-ltp6q\" (UID: \"77e59dd3-1218-4204-8573-24c3fa35d41b\") " pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.537692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e59dd3-1218-4204-8573-24c3fa35d41b-operator-scripts\") pod \"cinder-db-create-ltp6q\" (UID: \"77e59dd3-1218-4204-8573-24c3fa35d41b\") " pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.538554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e59dd3-1218-4204-8573-24c3fa35d41b-operator-scripts\") pod \"cinder-db-create-ltp6q\" (UID: \"77e59dd3-1218-4204-8573-24c3fa35d41b\") " pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.553064 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-76b8-account-create-945pv"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.554478 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.563379 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.575461 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nzwdl"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.589619 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-76b8-account-create-945pv"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.607993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v822r\" (UniqueName: \"kubernetes.io/projected/77e59dd3-1218-4204-8573-24c3fa35d41b-kube-api-access-v822r\") pod \"cinder-db-create-ltp6q\" (UID: \"77e59dd3-1218-4204-8573-24c3fa35d41b\") " pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.642024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnwsh\" (UniqueName: \"kubernetes.io/projected/945888ec-9f85-407e-b7d1-1feaac5c5e48-kube-api-access-rnwsh\") pod \"barbican-76b8-account-create-945pv\" (UID: \"945888ec-9f85-407e-b7d1-1feaac5c5e48\") " pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.642736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/945888ec-9f85-407e-b7d1-1feaac5c5e48-operator-scripts\") pod \"barbican-76b8-account-create-945pv\" (UID: \"945888ec-9f85-407e-b7d1-1feaac5c5e48\") " pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.724968 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.744751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/945888ec-9f85-407e-b7d1-1feaac5c5e48-operator-scripts\") pod \"barbican-76b8-account-create-945pv\" (UID: \"945888ec-9f85-407e-b7d1-1feaac5c5e48\") " pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.744883 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phtdj\" (UniqueName: \"kubernetes.io/projected/38a48af4-2b60-429d-9037-e483a357cde3-kube-api-access-phtdj\") pod \"barbican-db-create-nzwdl\" (UID: \"38a48af4-2b60-429d-9037-e483a357cde3\") " pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.744918 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a48af4-2b60-429d-9037-e483a357cde3-operator-scripts\") pod \"barbican-db-create-nzwdl\" (UID: \"38a48af4-2b60-429d-9037-e483a357cde3\") " pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.744962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnwsh\" (UniqueName: \"kubernetes.io/projected/945888ec-9f85-407e-b7d1-1feaac5c5e48-kube-api-access-rnwsh\") pod \"barbican-76b8-account-create-945pv\" (UID: \"945888ec-9f85-407e-b7d1-1feaac5c5e48\") " pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.745943 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/945888ec-9f85-407e-b7d1-1feaac5c5e48-operator-scripts\") pod \"barbican-76b8-account-create-945pv\" (UID: \"945888ec-9f85-407e-b7d1-1feaac5c5e48\") " pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.764257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.776198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnwsh\" (UniqueName: \"kubernetes.io/projected/945888ec-9f85-407e-b7d1-1feaac5c5e48-kube-api-access-rnwsh\") pod \"barbican-76b8-account-create-945pv\" (UID: \"945888ec-9f85-407e-b7d1-1feaac5c5e48\") " pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.810872 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-829a-account-create-zt6bp"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.811869 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.817364 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.846005 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phtdj\" (UniqueName: \"kubernetes.io/projected/38a48af4-2b60-429d-9037-e483a357cde3-kube-api-access-phtdj\") pod \"barbican-db-create-nzwdl\" (UID: \"38a48af4-2b60-429d-9037-e483a357cde3\") " pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.846092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a48af4-2b60-429d-9037-e483a357cde3-operator-scripts\") pod \"barbican-db-create-nzwdl\" (UID: \"38a48af4-2b60-429d-9037-e483a357cde3\") " pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.847758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a48af4-2b60-429d-9037-e483a357cde3-operator-scripts\") pod \"barbican-db-create-nzwdl\" (UID: \"38a48af4-2b60-429d-9037-e483a357cde3\") " pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.857340 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-829a-account-create-zt6bp"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.877178 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.882362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phtdj\" (UniqueName: \"kubernetes.io/projected/38a48af4-2b60-429d-9037-e483a357cde3-kube-api-access-phtdj\") pod \"barbican-db-create-nzwdl\" (UID: \"38a48af4-2b60-429d-9037-e483a357cde3\") " pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.891456 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-lp9lk"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.892609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.926706 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lp9lk"] Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.949883 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c0a2dfb-5b32-409d-9f07-87af1957f886-operator-scripts\") pod \"cinder-829a-account-create-zt6bp\" (UID: \"3c0a2dfb-5b32-409d-9f07-87af1957f886\") " pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:18 crc kubenswrapper[4824]: I1124 13:33:18.949949 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxrpj\" (UniqueName: \"kubernetes.io/projected/3c0a2dfb-5b32-409d-9f07-87af1957f886-kube-api-access-jxrpj\") pod \"cinder-829a-account-create-zt6bp\" (UID: \"3c0a2dfb-5b32-409d-9f07-87af1957f886\") " pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.042322 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-4b88-account-create-ld6br"] Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.043397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.048250 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.054752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxrpj\" (UniqueName: \"kubernetes.io/projected/3c0a2dfb-5b32-409d-9f07-87af1957f886-kube-api-access-jxrpj\") pod \"cinder-829a-account-create-zt6bp\" (UID: \"3c0a2dfb-5b32-409d-9f07-87af1957f886\") " pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.054862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af46b974-486a-4b7c-8102-1c70f023cac2-operator-scripts\") pod \"neutron-db-create-lp9lk\" (UID: \"af46b974-486a-4b7c-8102-1c70f023cac2\") " pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.054920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhszl\" (UniqueName: \"kubernetes.io/projected/af46b974-486a-4b7c-8102-1c70f023cac2-kube-api-access-dhszl\") pod \"neutron-db-create-lp9lk\" (UID: \"af46b974-486a-4b7c-8102-1c70f023cac2\") " pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.054957 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c0a2dfb-5b32-409d-9f07-87af1957f886-operator-scripts\") pod \"cinder-829a-account-create-zt6bp\" (UID: \"3c0a2dfb-5b32-409d-9f07-87af1957f886\") " pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.055539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c0a2dfb-5b32-409d-9f07-87af1957f886-operator-scripts\") pod \"cinder-829a-account-create-zt6bp\" (UID: \"3c0a2dfb-5b32-409d-9f07-87af1957f886\") " pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.056031 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-4b88-account-create-ld6br"] Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.093977 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-x4w8p"] Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.095061 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.101488 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.101599 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.101844 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k9jvt" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.101953 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.103678 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxrpj\" (UniqueName: \"kubernetes.io/projected/3c0a2dfb-5b32-409d-9f07-87af1957f886-kube-api-access-jxrpj\") pod \"cinder-829a-account-create-zt6bp\" (UID: \"3c0a2dfb-5b32-409d-9f07-87af1957f886\") " pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.136893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-x4w8p"] Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.137873 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.157611 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f77f534-a41d-46dc-9ac6-faf212fed300-operator-scripts\") pod \"neutron-4b88-account-create-ld6br\" (UID: \"6f77f534-a41d-46dc-9ac6-faf212fed300\") " pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.157667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhszl\" (UniqueName: \"kubernetes.io/projected/af46b974-486a-4b7c-8102-1c70f023cac2-kube-api-access-dhszl\") pod \"neutron-db-create-lp9lk\" (UID: \"af46b974-486a-4b7c-8102-1c70f023cac2\") " pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.157711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rbww\" (UniqueName: \"kubernetes.io/projected/6f77f534-a41d-46dc-9ac6-faf212fed300-kube-api-access-8rbww\") pod \"neutron-4b88-account-create-ld6br\" (UID: \"6f77f534-a41d-46dc-9ac6-faf212fed300\") " pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.157777 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af46b974-486a-4b7c-8102-1c70f023cac2-operator-scripts\") pod \"neutron-db-create-lp9lk\" (UID: \"af46b974-486a-4b7c-8102-1c70f023cac2\") " pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.158639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af46b974-486a-4b7c-8102-1c70f023cac2-operator-scripts\") pod \"neutron-db-create-lp9lk\" (UID: \"af46b974-486a-4b7c-8102-1c70f023cac2\") " pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.159859 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.176948 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhszl\" (UniqueName: \"kubernetes.io/projected/af46b974-486a-4b7c-8102-1c70f023cac2-kube-api-access-dhszl\") pod \"neutron-db-create-lp9lk\" (UID: \"af46b974-486a-4b7c-8102-1c70f023cac2\") " pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.232277 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.259602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f77f534-a41d-46dc-9ac6-faf212fed300-operator-scripts\") pod \"neutron-4b88-account-create-ld6br\" (UID: \"6f77f534-a41d-46dc-9ac6-faf212fed300\") " pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.259859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bk5j\" (UniqueName: \"kubernetes.io/projected/13797045-20bc-4f20-bf13-47b71e2f3342-kube-api-access-9bk5j\") pod \"keystone-db-sync-x4w8p\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.260001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rbww\" (UniqueName: \"kubernetes.io/projected/6f77f534-a41d-46dc-9ac6-faf212fed300-kube-api-access-8rbww\") pod \"neutron-4b88-account-create-ld6br\" (UID: \"6f77f534-a41d-46dc-9ac6-faf212fed300\") " pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.260184 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-combined-ca-bundle\") pod \"keystone-db-sync-x4w8p\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.260301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-config-data\") pod \"keystone-db-sync-x4w8p\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.260377 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f77f534-a41d-46dc-9ac6-faf212fed300-operator-scripts\") pod \"neutron-4b88-account-create-ld6br\" (UID: \"6f77f534-a41d-46dc-9ac6-faf212fed300\") " pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.280370 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rbww\" (UniqueName: \"kubernetes.io/projected/6f77f534-a41d-46dc-9ac6-faf212fed300-kube-api-access-8rbww\") pod \"neutron-4b88-account-create-ld6br\" (UID: \"6f77f534-a41d-46dc-9ac6-faf212fed300\") " pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.362465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-combined-ca-bundle\") pod \"keystone-db-sync-x4w8p\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.362555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-config-data\") pod \"keystone-db-sync-x4w8p\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.362614 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bk5j\" (UniqueName: \"kubernetes.io/projected/13797045-20bc-4f20-bf13-47b71e2f3342-kube-api-access-9bk5j\") pod \"keystone-db-sync-x4w8p\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.366716 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-combined-ca-bundle\") pod \"keystone-db-sync-x4w8p\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.367242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-config-data\") pod \"keystone-db-sync-x4w8p\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.377492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.385281 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bk5j\" (UniqueName: \"kubernetes.io/projected/13797045-20bc-4f20-bf13-47b71e2f3342-kube-api-access-9bk5j\") pod \"keystone-db-sync-x4w8p\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:19 crc kubenswrapper[4824]: I1124 13:33:19.439446 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.370835 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-x4w8p"] Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.388101 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ds8fs-config-s9ngk"] Nov 24 13:33:21 crc kubenswrapper[4824]: W1124 13:33:21.408541 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13797045_20bc_4f20_bf13_47b71e2f3342.slice/crio-87d0a4e54feccffe27f17fcdb163bab79a6246f8277baa050f5729e475f91b96 WatchSource:0}: Error finding container 87d0a4e54feccffe27f17fcdb163bab79a6246f8277baa050f5729e475f91b96: Status 404 returned error can't find the container with id 87d0a4e54feccffe27f17fcdb163bab79a6246f8277baa050f5729e475f91b96 Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.681926 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lp9lk"] Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.692365 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-4b88-account-create-ld6br"] Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.711001 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nzwdl"] Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.717662 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ltp6q"] Nov 24 13:33:21 crc kubenswrapper[4824]: W1124 13:33:21.761305 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f77f534_a41d_46dc_9ac6_faf212fed300.slice/crio-97adf38632539ff6bf8f2fcde8e4bf5e3ca211d15c742ce549ac178cf01457d4 WatchSource:0}: Error finding container 97adf38632539ff6bf8f2fcde8e4bf5e3ca211d15c742ce549ac178cf01457d4: Status 404 returned error can't find the container with id 97adf38632539ff6bf8f2fcde8e4bf5e3ca211d15c742ce549ac178cf01457d4 Nov 24 13:33:21 crc kubenswrapper[4824]: W1124 13:33:21.772084 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf46b974_486a_4b7c_8102_1c70f023cac2.slice/crio-caa21954a1e9f35f5ee4f232330166b56e2625a800c6eb13158967252a017709 WatchSource:0}: Error finding container caa21954a1e9f35f5ee4f232330166b56e2625a800c6eb13158967252a017709: Status 404 returned error can't find the container with id caa21954a1e9f35f5ee4f232330166b56e2625a800c6eb13158967252a017709 Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.817030 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-76b8-account-create-945pv"] Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.878258 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-829a-account-create-zt6bp"] Nov 24 13:33:21 crc kubenswrapper[4824]: W1124 13:33:21.880044 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77e59dd3_1218_4204_8573_24c3fa35d41b.slice/crio-1dc0df26fdae824d4c492477821eee5f9f526512431423e4e29682851a4ccde8 WatchSource:0}: Error finding container 1dc0df26fdae824d4c492477821eee5f9f526512431423e4e29682851a4ccde8: Status 404 returned error can't find the container with id 1dc0df26fdae824d4c492477821eee5f9f526512431423e4e29682851a4ccde8 Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.907248 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 13:33:21 crc kubenswrapper[4824]: I1124 13:33:21.936839 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.254761 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-x4w8p" event={"ID":"13797045-20bc-4f20-bf13-47b71e2f3342","Type":"ContainerStarted","Data":"87d0a4e54feccffe27f17fcdb163bab79a6246f8277baa050f5729e475f91b96"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.256540 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ds8fs-config-s9ngk" event={"ID":"5d043533-d754-4424-937a-958f424b54dd","Type":"ContainerStarted","Data":"260fbee5605e0180a413e46b64f14f0374ba776b4d5d74aada3ce1635c31a16b"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.256581 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ds8fs-config-s9ngk" event={"ID":"5d043533-d754-4424-937a-958f424b54dd","Type":"ContainerStarted","Data":"c80042675f9ff5665521d029e3eefe7eb7305d2c1738f87a090603bbe9a0d29f"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.261264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-76b8-account-create-945pv" event={"ID":"945888ec-9f85-407e-b7d1-1feaac5c5e48","Type":"ContainerStarted","Data":"313569d67d4dd322ca7cac67b7de73b5b96063d622399db405bb55a5abffaabe"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.262372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ltp6q" event={"ID":"77e59dd3-1218-4204-8573-24c3fa35d41b","Type":"ContainerStarted","Data":"1dc0df26fdae824d4c492477821eee5f9f526512431423e4e29682851a4ccde8"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.263249 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nzwdl" event={"ID":"38a48af4-2b60-429d-9037-e483a357cde3","Type":"ContainerStarted","Data":"fda36b1fd7746549a1b45677f709b03aef2cd58ab175722a0cd9a5f04a75007a"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.264488 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4b88-account-create-ld6br" event={"ID":"6f77f534-a41d-46dc-9ac6-faf212fed300","Type":"ContainerStarted","Data":"9c615a9cf2ff228c5f8bbb889943d717da8081ca0eaae2ba5b6be1b6cb28445a"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.264513 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4b88-account-create-ld6br" event={"ID":"6f77f534-a41d-46dc-9ac6-faf212fed300","Type":"ContainerStarted","Data":"97adf38632539ff6bf8f2fcde8e4bf5e3ca211d15c742ce549ac178cf01457d4"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.267710 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-829a-account-create-zt6bp" event={"ID":"3c0a2dfb-5b32-409d-9f07-87af1957f886","Type":"ContainerStarted","Data":"31cf2cbdcc640621c33408f207857c5e31c02f518793f81c2d5db9964a10aa17"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.276514 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ds8fs" podUID="b3c188e2-9a89-47fe-8efe-1187a3c81552" containerName="ovn-controller" probeResult="failure" output=< Nov 24 13:33:22 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 13:33:22 crc kubenswrapper[4824]: > Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.276626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"b8afcc287ca8fc1ec48406f36a9de6723d5079ea99b993fe8c6c8479ac08c42b"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.276663 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"666edb0f506a18afe4446ae6ff4bd1d9341948fda2563e6c56a7736bf61c4fbd"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.281456 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ds8fs-config-s9ngk" podStartSLOduration=10.281441218 podStartE2EDuration="10.281441218s" podCreationTimestamp="2025-11-24 13:33:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:22.277015954 +0000 UTC m=+1143.916555254" watchObservedRunningTime="2025-11-24 13:33:22.281441218 +0000 UTC m=+1143.920980518" Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.284710 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sb5fc" event={"ID":"0b623f0c-385a-465d-9132-9406dfdec4ec","Type":"ContainerStarted","Data":"3235005020b9c2b3de53baed0de55b2479d2b7cb39773877fd34bc2c12a625ef"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.287359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lp9lk" event={"ID":"af46b974-486a-4b7c-8102-1c70f023cac2","Type":"ContainerStarted","Data":"caa21954a1e9f35f5ee4f232330166b56e2625a800c6eb13158967252a017709"} Nov 24 13:33:22 crc kubenswrapper[4824]: I1124 13:33:22.293692 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-4b88-account-create-ld6br" podStartSLOduration=3.293673853 podStartE2EDuration="3.293673853s" podCreationTimestamp="2025-11-24 13:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:22.292773849 +0000 UTC m=+1143.932313159" watchObservedRunningTime="2025-11-24 13:33:22.293673853 +0000 UTC m=+1143.933213163" Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.326523 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"8111d48bd75a930c3f28d06a21b8e83b260f46baaa234444eb74bb9ea4bae667"} Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.326933 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"739ba7199374b1cc15c70ef9fdd032c4cde80478ff04a48eeed9d087f6a71951"} Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.332008 4824 generic.go:334] "Generic (PLEG): container finished" podID="77e59dd3-1218-4204-8573-24c3fa35d41b" containerID="46993937eaf72dd5104a840a6466f2d6efeddd43a101453f90366c039138f1b9" exitCode=0 Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.332191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ltp6q" event={"ID":"77e59dd3-1218-4204-8573-24c3fa35d41b","Type":"ContainerDied","Data":"46993937eaf72dd5104a840a6466f2d6efeddd43a101453f90366c039138f1b9"} Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.339674 4824 generic.go:334] "Generic (PLEG): container finished" podID="38a48af4-2b60-429d-9037-e483a357cde3" containerID="a0f8b224e94e354de91ffe9c2ed72b311c9afd81a58cc6d65e1f257ea59e8c42" exitCode=0 Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.339751 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nzwdl" event={"ID":"38a48af4-2b60-429d-9037-e483a357cde3","Type":"ContainerDied","Data":"a0f8b224e94e354de91ffe9c2ed72b311c9afd81a58cc6d65e1f257ea59e8c42"} Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.341142 4824 generic.go:334] "Generic (PLEG): container finished" podID="af46b974-486a-4b7c-8102-1c70f023cac2" containerID="d1f53bd96f5a5dcc7e5c904637aeed92816ecb28bbb1dfef4a00c402c27794df" exitCode=0 Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.341192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lp9lk" event={"ID":"af46b974-486a-4b7c-8102-1c70f023cac2","Type":"ContainerDied","Data":"d1f53bd96f5a5dcc7e5c904637aeed92816ecb28bbb1dfef4a00c402c27794df"} Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.346278 4824 generic.go:334] "Generic (PLEG): container finished" podID="6f77f534-a41d-46dc-9ac6-faf212fed300" containerID="9c615a9cf2ff228c5f8bbb889943d717da8081ca0eaae2ba5b6be1b6cb28445a" exitCode=0 Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.346622 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4b88-account-create-ld6br" event={"ID":"6f77f534-a41d-46dc-9ac6-faf212fed300","Type":"ContainerDied","Data":"9c615a9cf2ff228c5f8bbb889943d717da8081ca0eaae2ba5b6be1b6cb28445a"} Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.349864 4824 generic.go:334] "Generic (PLEG): container finished" podID="3c0a2dfb-5b32-409d-9f07-87af1957f886" containerID="5bd31d9784253bd14a5bdc2e681826c9955f8928e1df547da0f6faee12c815dc" exitCode=0 Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.349950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-829a-account-create-zt6bp" event={"ID":"3c0a2dfb-5b32-409d-9f07-87af1957f886","Type":"ContainerDied","Data":"5bd31d9784253bd14a5bdc2e681826c9955f8928e1df547da0f6faee12c815dc"} Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.353073 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-sb5fc" podStartSLOduration=3.639850152 podStartE2EDuration="16.353044732s" podCreationTimestamp="2025-11-24 13:33:07 +0000 UTC" firstStartedPulling="2025-11-24 13:33:08.071506665 +0000 UTC m=+1129.711045975" lastFinishedPulling="2025-11-24 13:33:20.784701245 +0000 UTC m=+1142.424240555" observedRunningTime="2025-11-24 13:33:22.311351288 +0000 UTC m=+1143.950890588" watchObservedRunningTime="2025-11-24 13:33:23.353044732 +0000 UTC m=+1144.992584102" Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.359760 4824 generic.go:334] "Generic (PLEG): container finished" podID="5d043533-d754-4424-937a-958f424b54dd" containerID="260fbee5605e0180a413e46b64f14f0374ba776b4d5d74aada3ce1635c31a16b" exitCode=0 Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.359934 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ds8fs-config-s9ngk" event={"ID":"5d043533-d754-4424-937a-958f424b54dd","Type":"ContainerDied","Data":"260fbee5605e0180a413e46b64f14f0374ba776b4d5d74aada3ce1635c31a16b"} Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.361904 4824 generic.go:334] "Generic (PLEG): container finished" podID="945888ec-9f85-407e-b7d1-1feaac5c5e48" containerID="f4730305e6c94d84d1983d5c4b44d3b9ba745e9575bf6bad94a7cd36271308f9" exitCode=0 Nov 24 13:33:23 crc kubenswrapper[4824]: I1124 13:33:23.362401 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-76b8-account-create-945pv" event={"ID":"945888ec-9f85-407e-b7d1-1feaac5c5e48","Type":"ContainerDied","Data":"f4730305e6c94d84d1983d5c4b44d3b9ba745e9575bf6bad94a7cd36271308f9"} Nov 24 13:33:27 crc kubenswrapper[4824]: I1124 13:33:27.185095 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ds8fs" Nov 24 13:33:27 crc kubenswrapper[4824]: I1124 13:33:27.950429 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:27 crc kubenswrapper[4824]: I1124 13:33:27.989071 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.003195 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.037149 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.049440 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.066735 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmkls\" (UniqueName: \"kubernetes.io/projected/5d043533-d754-4424-937a-958f424b54dd-kube-api-access-tmkls\") pod \"5d043533-d754-4424-937a-958f424b54dd\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.066788 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run\") pod \"5d043533-d754-4424-937a-958f424b54dd\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.066833 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxrpj\" (UniqueName: \"kubernetes.io/projected/3c0a2dfb-5b32-409d-9f07-87af1957f886-kube-api-access-jxrpj\") pod \"3c0a2dfb-5b32-409d-9f07-87af1957f886\" (UID: \"3c0a2dfb-5b32-409d-9f07-87af1957f886\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.066859 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-log-ovn\") pod \"5d043533-d754-4424-937a-958f424b54dd\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.066893 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/945888ec-9f85-407e-b7d1-1feaac5c5e48-operator-scripts\") pod \"945888ec-9f85-407e-b7d1-1feaac5c5e48\" (UID: \"945888ec-9f85-407e-b7d1-1feaac5c5e48\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.066927 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e59dd3-1218-4204-8573-24c3fa35d41b-operator-scripts\") pod \"77e59dd3-1218-4204-8573-24c3fa35d41b\" (UID: \"77e59dd3-1218-4204-8573-24c3fa35d41b\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.066954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c0a2dfb-5b32-409d-9f07-87af1957f886-operator-scripts\") pod \"3c0a2dfb-5b32-409d-9f07-87af1957f886\" (UID: \"3c0a2dfb-5b32-409d-9f07-87af1957f886\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.066978 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f77f534-a41d-46dc-9ac6-faf212fed300-operator-scripts\") pod \"6f77f534-a41d-46dc-9ac6-faf212fed300\" (UID: \"6f77f534-a41d-46dc-9ac6-faf212fed300\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.067019 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run-ovn\") pod \"5d043533-d754-4424-937a-958f424b54dd\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.067045 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnwsh\" (UniqueName: \"kubernetes.io/projected/945888ec-9f85-407e-b7d1-1feaac5c5e48-kube-api-access-rnwsh\") pod \"945888ec-9f85-407e-b7d1-1feaac5c5e48\" (UID: \"945888ec-9f85-407e-b7d1-1feaac5c5e48\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.067058 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "5d043533-d754-4424-937a-958f424b54dd" (UID: "5d043533-d754-4424-937a-958f424b54dd"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.067126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-scripts\") pod \"5d043533-d754-4424-937a-958f424b54dd\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.067155 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v822r\" (UniqueName: \"kubernetes.io/projected/77e59dd3-1218-4204-8573-24c3fa35d41b-kube-api-access-v822r\") pod \"77e59dd3-1218-4204-8573-24c3fa35d41b\" (UID: \"77e59dd3-1218-4204-8573-24c3fa35d41b\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.067179 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-additional-scripts\") pod \"5d043533-d754-4424-937a-958f424b54dd\" (UID: \"5d043533-d754-4424-937a-958f424b54dd\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.067228 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rbww\" (UniqueName: \"kubernetes.io/projected/6f77f534-a41d-46dc-9ac6-faf212fed300-kube-api-access-8rbww\") pod \"6f77f534-a41d-46dc-9ac6-faf212fed300\" (UID: \"6f77f534-a41d-46dc-9ac6-faf212fed300\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.067491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run" (OuterVolumeSpecName: "var-run") pod "5d043533-d754-4424-937a-958f424b54dd" (UID: "5d043533-d754-4424-937a-958f424b54dd"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.067624 4824 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.068754 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/945888ec-9f85-407e-b7d1-1feaac5c5e48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "945888ec-9f85-407e-b7d1-1feaac5c5e48" (UID: "945888ec-9f85-407e-b7d1-1feaac5c5e48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.069710 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-scripts" (OuterVolumeSpecName: "scripts") pod "5d043533-d754-4424-937a-958f424b54dd" (UID: "5d043533-d754-4424-937a-958f424b54dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.070643 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c0a2dfb-5b32-409d-9f07-87af1957f886-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c0a2dfb-5b32-409d-9f07-87af1957f886" (UID: "3c0a2dfb-5b32-409d-9f07-87af1957f886"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.070917 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f77f534-a41d-46dc-9ac6-faf212fed300-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6f77f534-a41d-46dc-9ac6-faf212fed300" (UID: "6f77f534-a41d-46dc-9ac6-faf212fed300"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.071344 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.071700 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77e59dd3-1218-4204-8573-24c3fa35d41b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77e59dd3-1218-4204-8573-24c3fa35d41b" (UID: "77e59dd3-1218-4204-8573-24c3fa35d41b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.072143 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "5d043533-d754-4424-937a-958f424b54dd" (UID: "5d043533-d754-4424-937a-958f424b54dd"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.072266 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "5d043533-d754-4424-937a-958f424b54dd" (UID: "5d043533-d754-4424-937a-958f424b54dd"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.077411 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d043533-d754-4424-937a-958f424b54dd-kube-api-access-tmkls" (OuterVolumeSpecName: "kube-api-access-tmkls") pod "5d043533-d754-4424-937a-958f424b54dd" (UID: "5d043533-d754-4424-937a-958f424b54dd"). InnerVolumeSpecName "kube-api-access-tmkls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.080743 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77e59dd3-1218-4204-8573-24c3fa35d41b-kube-api-access-v822r" (OuterVolumeSpecName: "kube-api-access-v822r") pod "77e59dd3-1218-4204-8573-24c3fa35d41b" (UID: "77e59dd3-1218-4204-8573-24c3fa35d41b"). InnerVolumeSpecName "kube-api-access-v822r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.081362 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f77f534-a41d-46dc-9ac6-faf212fed300-kube-api-access-8rbww" (OuterVolumeSpecName: "kube-api-access-8rbww") pod "6f77f534-a41d-46dc-9ac6-faf212fed300" (UID: "6f77f534-a41d-46dc-9ac6-faf212fed300"). InnerVolumeSpecName "kube-api-access-8rbww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.082454 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/945888ec-9f85-407e-b7d1-1feaac5c5e48-kube-api-access-rnwsh" (OuterVolumeSpecName: "kube-api-access-rnwsh") pod "945888ec-9f85-407e-b7d1-1feaac5c5e48" (UID: "945888ec-9f85-407e-b7d1-1feaac5c5e48"). InnerVolumeSpecName "kube-api-access-rnwsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.082579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c0a2dfb-5b32-409d-9f07-87af1957f886-kube-api-access-jxrpj" (OuterVolumeSpecName: "kube-api-access-jxrpj") pod "3c0a2dfb-5b32-409d-9f07-87af1957f886" (UID: "3c0a2dfb-5b32-409d-9f07-87af1957f886"). InnerVolumeSpecName "kube-api-access-jxrpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.093186 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.168500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a48af4-2b60-429d-9037-e483a357cde3-operator-scripts\") pod \"38a48af4-2b60-429d-9037-e483a357cde3\" (UID: \"38a48af4-2b60-429d-9037-e483a357cde3\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.168547 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af46b974-486a-4b7c-8102-1c70f023cac2-operator-scripts\") pod \"af46b974-486a-4b7c-8102-1c70f023cac2\" (UID: \"af46b974-486a-4b7c-8102-1c70f023cac2\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.168954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phtdj\" (UniqueName: \"kubernetes.io/projected/38a48af4-2b60-429d-9037-e483a357cde3-kube-api-access-phtdj\") pod \"38a48af4-2b60-429d-9037-e483a357cde3\" (UID: \"38a48af4-2b60-429d-9037-e483a357cde3\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169032 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhszl\" (UniqueName: \"kubernetes.io/projected/af46b974-486a-4b7c-8102-1c70f023cac2-kube-api-access-dhszl\") pod \"af46b974-486a-4b7c-8102-1c70f023cac2\" (UID: \"af46b974-486a-4b7c-8102-1c70f023cac2\") " Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169475 4824 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169501 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxrpj\" (UniqueName: \"kubernetes.io/projected/3c0a2dfb-5b32-409d-9f07-87af1957f886-kube-api-access-jxrpj\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169517 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/945888ec-9f85-407e-b7d1-1feaac5c5e48-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169531 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e59dd3-1218-4204-8573-24c3fa35d41b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169542 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c0a2dfb-5b32-409d-9f07-87af1957f886-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169553 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f77f534-a41d-46dc-9ac6-faf212fed300-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169565 4824 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5d043533-d754-4424-937a-958f424b54dd-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169577 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnwsh\" (UniqueName: \"kubernetes.io/projected/945888ec-9f85-407e-b7d1-1feaac5c5e48-kube-api-access-rnwsh\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169590 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169602 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v822r\" (UniqueName: \"kubernetes.io/projected/77e59dd3-1218-4204-8573-24c3fa35d41b-kube-api-access-v822r\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169614 4824 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5d043533-d754-4424-937a-958f424b54dd-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169630 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rbww\" (UniqueName: \"kubernetes.io/projected/6f77f534-a41d-46dc-9ac6-faf212fed300-kube-api-access-8rbww\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.169643 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmkls\" (UniqueName: \"kubernetes.io/projected/5d043533-d754-4424-937a-958f424b54dd-kube-api-access-tmkls\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.170521 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af46b974-486a-4b7c-8102-1c70f023cac2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af46b974-486a-4b7c-8102-1c70f023cac2" (UID: "af46b974-486a-4b7c-8102-1c70f023cac2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.171169 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38a48af4-2b60-429d-9037-e483a357cde3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "38a48af4-2b60-429d-9037-e483a357cde3" (UID: "38a48af4-2b60-429d-9037-e483a357cde3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.173052 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af46b974-486a-4b7c-8102-1c70f023cac2-kube-api-access-dhszl" (OuterVolumeSpecName: "kube-api-access-dhszl") pod "af46b974-486a-4b7c-8102-1c70f023cac2" (UID: "af46b974-486a-4b7c-8102-1c70f023cac2"). InnerVolumeSpecName "kube-api-access-dhszl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.173120 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38a48af4-2b60-429d-9037-e483a357cde3-kube-api-access-phtdj" (OuterVolumeSpecName: "kube-api-access-phtdj") pod "38a48af4-2b60-429d-9037-e483a357cde3" (UID: "38a48af4-2b60-429d-9037-e483a357cde3"). InnerVolumeSpecName "kube-api-access-phtdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.270401 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a48af4-2b60-429d-9037-e483a357cde3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.270429 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af46b974-486a-4b7c-8102-1c70f023cac2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.270438 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phtdj\" (UniqueName: \"kubernetes.io/projected/38a48af4-2b60-429d-9037-e483a357cde3-kube-api-access-phtdj\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.270449 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhszl\" (UniqueName: \"kubernetes.io/projected/af46b974-486a-4b7c-8102-1c70f023cac2-kube-api-access-dhszl\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.419227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-x4w8p" event={"ID":"13797045-20bc-4f20-bf13-47b71e2f3342","Type":"ContainerStarted","Data":"23c2b8f40274336e54fad1534e72aeb73219e9b1b9067768f4767000eea23a50"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.421628 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-829a-account-create-zt6bp" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.422485 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-829a-account-create-zt6bp" event={"ID":"3c0a2dfb-5b32-409d-9f07-87af1957f886","Type":"ContainerDied","Data":"31cf2cbdcc640621c33408f207857c5e31c02f518793f81c2d5db9964a10aa17"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.422521 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31cf2cbdcc640621c33408f207857c5e31c02f518793f81c2d5db9964a10aa17" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.445054 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ds8fs-config-s9ngk" event={"ID":"5d043533-d754-4424-937a-958f424b54dd","Type":"ContainerDied","Data":"c80042675f9ff5665521d029e3eefe7eb7305d2c1738f87a090603bbe9a0d29f"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.445093 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c80042675f9ff5665521d029e3eefe7eb7305d2c1738f87a090603bbe9a0d29f" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.445157 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ds8fs-config-s9ngk" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.455466 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-x4w8p" podStartSLOduration=3.101436002 podStartE2EDuration="9.455446931s" podCreationTimestamp="2025-11-24 13:33:19 +0000 UTC" firstStartedPulling="2025-11-24 13:33:21.410956052 +0000 UTC m=+1143.050495362" lastFinishedPulling="2025-11-24 13:33:27.764966981 +0000 UTC m=+1149.404506291" observedRunningTime="2025-11-24 13:33:28.43791962 +0000 UTC m=+1150.077458930" watchObservedRunningTime="2025-11-24 13:33:28.455446931 +0000 UTC m=+1150.094986251" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.475085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"44138bf4ecae02e54407869707623362afb0abecfe6e90a5f1e9288c42ff7f8b"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.475218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"d984b32cfbf8aa812331e6511a1f128af778f8af42374a632e473275cd766850"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.475237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"4a61299faa96ae25a768b33c226e13f5f6b332b2e9efa97491f457cd0ef0e68b"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.485364 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-76b8-account-create-945pv" event={"ID":"945888ec-9f85-407e-b7d1-1feaac5c5e48","Type":"ContainerDied","Data":"313569d67d4dd322ca7cac67b7de73b5b96063d622399db405bb55a5abffaabe"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.485439 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="313569d67d4dd322ca7cac67b7de73b5b96063d622399db405bb55a5abffaabe" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.485543 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-76b8-account-create-945pv" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.493516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ltp6q" event={"ID":"77e59dd3-1218-4204-8573-24c3fa35d41b","Type":"ContainerDied","Data":"1dc0df26fdae824d4c492477821eee5f9f526512431423e4e29682851a4ccde8"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.493570 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dc0df26fdae824d4c492477821eee5f9f526512431423e4e29682851a4ccde8" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.493651 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ltp6q" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.508516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nzwdl" event={"ID":"38a48af4-2b60-429d-9037-e483a357cde3","Type":"ContainerDied","Data":"fda36b1fd7746549a1b45677f709b03aef2cd58ab175722a0cd9a5f04a75007a"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.508565 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fda36b1fd7746549a1b45677f709b03aef2cd58ab175722a0cd9a5f04a75007a" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.508658 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nzwdl" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.514499 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lp9lk" event={"ID":"af46b974-486a-4b7c-8102-1c70f023cac2","Type":"ContainerDied","Data":"caa21954a1e9f35f5ee4f232330166b56e2625a800c6eb13158967252a017709"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.514533 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caa21954a1e9f35f5ee4f232330166b56e2625a800c6eb13158967252a017709" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.514597 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lp9lk" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.542381 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4b88-account-create-ld6br" event={"ID":"6f77f534-a41d-46dc-9ac6-faf212fed300","Type":"ContainerDied","Data":"97adf38632539ff6bf8f2fcde8e4bf5e3ca211d15c742ce549ac178cf01457d4"} Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.542423 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97adf38632539ff6bf8f2fcde8e4bf5e3ca211d15c742ce549ac178cf01457d4" Nov 24 13:33:28 crc kubenswrapper[4824]: I1124 13:33:28.542500 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4b88-account-create-ld6br" Nov 24 13:33:29 crc kubenswrapper[4824]: I1124 13:33:29.140075 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ds8fs-config-s9ngk"] Nov 24 13:33:29 crc kubenswrapper[4824]: I1124 13:33:29.147984 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ds8fs-config-s9ngk"] Nov 24 13:33:29 crc kubenswrapper[4824]: I1124 13:33:29.556598 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"57c9e3110cd2213a3db30ef23e90bde8dd309d884c72a0e96d850a2920c1e88d"} Nov 24 13:33:31 crc kubenswrapper[4824]: I1124 13:33:31.017835 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d043533-d754-4424-937a-958f424b54dd" path="/var/lib/kubelet/pods/5d043533-d754-4424-937a-958f424b54dd/volumes" Nov 24 13:33:31 crc kubenswrapper[4824]: I1124 13:33:31.577452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"06a6cbd2b35a81a8eca38b4dd6ae1792d9a0907df605f9c19b60ba1741c4b71e"} Nov 24 13:33:31 crc kubenswrapper[4824]: I1124 13:33:31.577756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"0b906663634ce4a9beaab3e59d38cb9b45e15ceeb74a8ffdd24ae0ce440561cb"} Nov 24 13:33:31 crc kubenswrapper[4824]: I1124 13:33:31.577769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"c0065140705e3d7782aa2324b5360c49f03d0fe9815d103f660008773bbb8473"} Nov 24 13:33:31 crc kubenswrapper[4824]: I1124 13:33:31.577779 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"b8a96ed3b89a817061b64d7139f3c9440884eb32f09c6e0c7bd7d04f886f385c"} Nov 24 13:33:31 crc kubenswrapper[4824]: I1124 13:33:31.577788 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"25ad4d3ea5cb514daed68835cdb9c297f02f14660a8738abdeb00a182327ec6f"} Nov 24 13:33:31 crc kubenswrapper[4824]: I1124 13:33:31.577797 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"9c23b93b515629095517b91353ca0d573485e04592e9aea8ba87970b35561015"} Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.585720 4824 generic.go:334] "Generic (PLEG): container finished" podID="13797045-20bc-4f20-bf13-47b71e2f3342" containerID="23c2b8f40274336e54fad1534e72aeb73219e9b1b9067768f4767000eea23a50" exitCode=0 Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.585785 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-x4w8p" event={"ID":"13797045-20bc-4f20-bf13-47b71e2f3342","Type":"ContainerDied","Data":"23c2b8f40274336e54fad1534e72aeb73219e9b1b9067768f4767000eea23a50"} Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.593085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c","Type":"ContainerStarted","Data":"6c816a2e3d598e9ca4a0214bce8ef9d00f4459e593629f9ce84ab38fb8cfff4b"} Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.638259 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.874306019 podStartE2EDuration="39.63823771s" podCreationTimestamp="2025-11-24 13:32:53 +0000 UTC" firstStartedPulling="2025-11-24 13:33:11.66426715 +0000 UTC m=+1133.303806460" lastFinishedPulling="2025-11-24 13:33:30.428198841 +0000 UTC m=+1152.067738151" observedRunningTime="2025-11-24 13:33:32.635445709 +0000 UTC m=+1154.274985019" watchObservedRunningTime="2025-11-24 13:33:32.63823771 +0000 UTC m=+1154.277777010" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.885088 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dck68"] Nov 24 13:33:32 crc kubenswrapper[4824]: E1124 13:33:32.885757 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="945888ec-9f85-407e-b7d1-1feaac5c5e48" containerName="mariadb-account-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.885776 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="945888ec-9f85-407e-b7d1-1feaac5c5e48" containerName="mariadb-account-create" Nov 24 13:33:32 crc kubenswrapper[4824]: E1124 13:33:32.885822 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d043533-d754-4424-937a-958f424b54dd" containerName="ovn-config" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.885831 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d043533-d754-4424-937a-958f424b54dd" containerName="ovn-config" Nov 24 13:33:32 crc kubenswrapper[4824]: E1124 13:33:32.885862 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38a48af4-2b60-429d-9037-e483a357cde3" containerName="mariadb-database-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.885871 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="38a48af4-2b60-429d-9037-e483a357cde3" containerName="mariadb-database-create" Nov 24 13:33:32 crc kubenswrapper[4824]: E1124 13:33:32.885885 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f77f534-a41d-46dc-9ac6-faf212fed300" containerName="mariadb-account-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.885893 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f77f534-a41d-46dc-9ac6-faf212fed300" containerName="mariadb-account-create" Nov 24 13:33:32 crc kubenswrapper[4824]: E1124 13:33:32.885908 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0a2dfb-5b32-409d-9f07-87af1957f886" containerName="mariadb-account-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.885916 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0a2dfb-5b32-409d-9f07-87af1957f886" containerName="mariadb-account-create" Nov 24 13:33:32 crc kubenswrapper[4824]: E1124 13:33:32.885939 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af46b974-486a-4b7c-8102-1c70f023cac2" containerName="mariadb-database-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.885947 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="af46b974-486a-4b7c-8102-1c70f023cac2" containerName="mariadb-database-create" Nov 24 13:33:32 crc kubenswrapper[4824]: E1124 13:33:32.885957 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e59dd3-1218-4204-8573-24c3fa35d41b" containerName="mariadb-database-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.885965 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e59dd3-1218-4204-8573-24c3fa35d41b" containerName="mariadb-database-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.886231 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f77f534-a41d-46dc-9ac6-faf212fed300" containerName="mariadb-account-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.886259 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="af46b974-486a-4b7c-8102-1c70f023cac2" containerName="mariadb-database-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.886279 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="77e59dd3-1218-4204-8573-24c3fa35d41b" containerName="mariadb-database-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.886292 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="945888ec-9f85-407e-b7d1-1feaac5c5e48" containerName="mariadb-account-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.886308 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d043533-d754-4424-937a-958f424b54dd" containerName="ovn-config" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.886321 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="38a48af4-2b60-429d-9037-e483a357cde3" containerName="mariadb-database-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.886340 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c0a2dfb-5b32-409d-9f07-87af1957f886" containerName="mariadb-account-create" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.887362 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.892153 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 13:33:32 crc kubenswrapper[4824]: I1124 13:33:32.897325 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dck68"] Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.046299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.046359 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r97nc\" (UniqueName: \"kubernetes.io/projected/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-kube-api-access-r97nc\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.046392 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-config\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.046458 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-svc\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.046477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.046494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.148635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-config\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.149076 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-svc\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.149202 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.149333 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.149556 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-config\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.149560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.149640 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r97nc\" (UniqueName: \"kubernetes.io/projected/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-kube-api-access-r97nc\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.150042 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-svc\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.150684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.150955 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.151438 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.174677 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r97nc\" (UniqueName: \"kubernetes.io/projected/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-kube-api-access-r97nc\") pod \"dnsmasq-dns-764c5664d7-dck68\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.201554 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.600933 4824 generic.go:334] "Generic (PLEG): container finished" podID="0b623f0c-385a-465d-9132-9406dfdec4ec" containerID="3235005020b9c2b3de53baed0de55b2479d2b7cb39773877fd34bc2c12a625ef" exitCode=0 Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.601031 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sb5fc" event={"ID":"0b623f0c-385a-465d-9132-9406dfdec4ec","Type":"ContainerDied","Data":"3235005020b9c2b3de53baed0de55b2479d2b7cb39773877fd34bc2c12a625ef"} Nov 24 13:33:33 crc kubenswrapper[4824]: I1124 13:33:33.777187 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dck68"] Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.010358 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.070303 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-config-data\") pod \"13797045-20bc-4f20-bf13-47b71e2f3342\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.070738 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-combined-ca-bundle\") pod \"13797045-20bc-4f20-bf13-47b71e2f3342\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.070960 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bk5j\" (UniqueName: \"kubernetes.io/projected/13797045-20bc-4f20-bf13-47b71e2f3342-kube-api-access-9bk5j\") pod \"13797045-20bc-4f20-bf13-47b71e2f3342\" (UID: \"13797045-20bc-4f20-bf13-47b71e2f3342\") " Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.077395 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13797045-20bc-4f20-bf13-47b71e2f3342-kube-api-access-9bk5j" (OuterVolumeSpecName: "kube-api-access-9bk5j") pod "13797045-20bc-4f20-bf13-47b71e2f3342" (UID: "13797045-20bc-4f20-bf13-47b71e2f3342"). InnerVolumeSpecName "kube-api-access-9bk5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.083363 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bk5j\" (UniqueName: \"kubernetes.io/projected/13797045-20bc-4f20-bf13-47b71e2f3342-kube-api-access-9bk5j\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.102394 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13797045-20bc-4f20-bf13-47b71e2f3342" (UID: "13797045-20bc-4f20-bf13-47b71e2f3342"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.119993 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-config-data" (OuterVolumeSpecName: "config-data") pod "13797045-20bc-4f20-bf13-47b71e2f3342" (UID: "13797045-20bc-4f20-bf13-47b71e2f3342"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.185125 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.185157 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13797045-20bc-4f20-bf13-47b71e2f3342-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.611595 4824 generic.go:334] "Generic (PLEG): container finished" podID="ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" containerID="86b0dd4a57a4f860fb9d2ff2df5a29b0ec615e8bb84358debd2211cd17c64c48" exitCode=0 Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.611704 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dck68" event={"ID":"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4","Type":"ContainerDied","Data":"86b0dd4a57a4f860fb9d2ff2df5a29b0ec615e8bb84358debd2211cd17c64c48"} Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.612004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dck68" event={"ID":"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4","Type":"ContainerStarted","Data":"ca2dc56148a435952d950c6d93e4ce558c6a52b62de414c4aca8611e84385158"} Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.614721 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-x4w8p" event={"ID":"13797045-20bc-4f20-bf13-47b71e2f3342","Type":"ContainerDied","Data":"87d0a4e54feccffe27f17fcdb163bab79a6246f8277baa050f5729e475f91b96"} Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.614750 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87d0a4e54feccffe27f17fcdb163bab79a6246f8277baa050f5729e475f91b96" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.614751 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-x4w8p" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.851668 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-r2442"] Nov 24 13:33:34 crc kubenswrapper[4824]: E1124 13:33:34.852182 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13797045-20bc-4f20-bf13-47b71e2f3342" containerName="keystone-db-sync" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.852200 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13797045-20bc-4f20-bf13-47b71e2f3342" containerName="keystone-db-sync" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.852391 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13797045-20bc-4f20-bf13-47b71e2f3342" containerName="keystone-db-sync" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.853343 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.865293 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.865528 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k9jvt" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.865654 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.865765 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.865906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 13:33:34 crc kubenswrapper[4824]: I1124 13:33:34.909375 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r2442"] Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.002201 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-fernet-keys\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.002289 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-scripts\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.002376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-credential-keys\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.002420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-combined-ca-bundle\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.002448 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbtcg\" (UniqueName: \"kubernetes.io/projected/ae1d0586-93a9-4b78-8d56-9826b90e8f25-kube-api-access-rbtcg\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.002470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-config-data\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.051671 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dck68"] Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.104459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-credential-keys\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.104512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-combined-ca-bundle\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.104539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbtcg\" (UniqueName: \"kubernetes.io/projected/ae1d0586-93a9-4b78-8d56-9826b90e8f25-kube-api-access-rbtcg\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.104559 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-config-data\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.104589 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-fernet-keys\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.104623 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-scripts\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.124576 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-fernet-keys\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.127750 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-config-data\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.131681 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-combined-ca-bundle\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.134415 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-scripts\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.139926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-credential-keys\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.166645 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-85frw"] Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.167488 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbtcg\" (UniqueName: \"kubernetes.io/projected/ae1d0586-93a9-4b78-8d56-9826b90e8f25-kube-api-access-rbtcg\") pod \"keystone-bootstrap-r2442\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.168367 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:35 crc kubenswrapper[4824]: I1124 13:33:35.236007 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-85frw"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.248657 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-69zcz"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.254279 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.278793 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7pc9z" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.279070 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.279189 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.285457 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307728 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-scripts\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307760 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-combined-ca-bundle\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dpm7\" (UniqueName: \"kubernetes.io/projected/b1af3874-3566-4785-a5fe-13375aa658c3-kube-api-access-5dpm7\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307808 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-db-sync-config-data\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307833 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkqf9\" (UniqueName: \"kubernetes.io/projected/a4fa0482-6635-4253-b922-5251c70c79d7-kube-api-access-hkqf9\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1af3874-3566-4785-a5fe-13375aa658c3-etc-machine-id\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-config-data\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-svc\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.307967 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.308004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-config\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.342512 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-95tmx"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.424329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432464 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-config\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432559 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432593 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-scripts\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-combined-ca-bundle\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432658 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dpm7\" (UniqueName: \"kubernetes.io/projected/b1af3874-3566-4785-a5fe-13375aa658c3-kube-api-access-5dpm7\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-db-sync-config-data\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkqf9\" (UniqueName: \"kubernetes.io/projected/a4fa0482-6635-4253-b922-5251c70c79d7-kube-api-access-hkqf9\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.432807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1af3874-3566-4785-a5fe-13375aa658c3-etc-machine-id\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.461941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-config-data\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.462023 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-svc\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.462894 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-svc\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.501110 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.501333 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.501456 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fmdf7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.538373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-combined-ca-bundle\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.543561 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-config\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.544111 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.544606 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.553492 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.553865 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-69zcz"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.553964 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1af3874-3566-4785-a5fe-13375aa658c3-etc-machine-id\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.586312 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-config\") pod \"neutron-db-sync-95tmx\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.586500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-combined-ca-bundle\") pod \"neutron-db-sync-95tmx\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.586526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l59l\" (UniqueName: \"kubernetes.io/projected/8e88a010-ca80-44fc-bd45-d741e3acdc47-kube-api-access-5l59l\") pod \"neutron-db-sync-95tmx\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.598895 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-db-sync-config-data\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.600043 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-config-data\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.604791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-scripts\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.607402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dpm7\" (UniqueName: \"kubernetes.io/projected/b1af3874-3566-4785-a5fe-13375aa658c3-kube-api-access-5dpm7\") pod \"cinder-db-sync-69zcz\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.621268 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-69zcz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.649381 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-95tmx"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.657660 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkqf9\" (UniqueName: \"kubernetes.io/projected/a4fa0482-6635-4253-b922-5251c70c79d7-kube-api-access-hkqf9\") pod \"dnsmasq-dns-5959f8865f-85frw\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.671436 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6bccbd8f7f-rjdmz"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.672697 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.679063 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.681660 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.681687 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-ntcxr" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.682532 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.682690 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.683875 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sb5fc" event={"ID":"0b623f0c-385a-465d-9132-9406dfdec4ec","Type":"ContainerDied","Data":"a619220c703562e02dea0a0b3e6e158f17118d6c6755fb021dedbf9478ab3363"} Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.683910 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a619220c703562e02dea0a0b3e6e158f17118d6c6755fb021dedbf9478ab3363" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.687675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-config\") pod \"neutron-db-sync-95tmx\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.687758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-combined-ca-bundle\") pod \"neutron-db-sync-95tmx\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.687776 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l59l\" (UniqueName: \"kubernetes.io/projected/8e88a010-ca80-44fc-bd45-d741e3acdc47-kube-api-access-5l59l\") pod \"neutron-db-sync-95tmx\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.716435 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-config\") pod \"neutron-db-sync-95tmx\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.718387 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-8wqh7"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.718547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-combined-ca-bundle\") pod \"neutron-db-sync-95tmx\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: E1124 13:33:35.718766 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b623f0c-385a-465d-9132-9406dfdec4ec" containerName="glance-db-sync" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.718778 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b623f0c-385a-465d-9132-9406dfdec4ec" containerName="glance-db-sync" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.719072 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b623f0c-385a-465d-9132-9406dfdec4ec" containerName="glance-db-sync" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.719643 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.723427 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.724118 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pgbzs" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.724289 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.726248 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l59l\" (UniqueName: \"kubernetes.io/projected/8e88a010-ca80-44fc-bd45-d741e3acdc47-kube-api-access-5l59l\") pod \"neutron-db-sync-95tmx\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.750774 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-85frw"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.751753 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.764629 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bccbd8f7f-rjdmz"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.792879 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-combined-ca-bundle\") pod \"0b623f0c-385a-465d-9132-9406dfdec4ec\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.792929 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-db-sync-config-data\") pod \"0b623f0c-385a-465d-9132-9406dfdec4ec\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.793013 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-config-data\") pod \"0b623f0c-385a-465d-9132-9406dfdec4ec\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.793138 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4676c\" (UniqueName: \"kubernetes.io/projected/0b623f0c-385a-465d-9132-9406dfdec4ec-kube-api-access-4676c\") pod \"0b623f0c-385a-465d-9132-9406dfdec4ec\" (UID: \"0b623f0c-385a-465d-9132-9406dfdec4ec\") " Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.793428 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4022c6cb-7074-46e7-a366-7005553fa97e-horizon-secret-key\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.793464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-config-data\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.793931 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-scripts\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.793994 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4022c6cb-7074-46e7-a366-7005553fa97e-logs\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.794128 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd7mt\" (UniqueName: \"kubernetes.io/projected/4022c6cb-7074-46e7-a366-7005553fa97e-kube-api-access-pd7mt\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.809638 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.811491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b623f0c-385a-465d-9132-9406dfdec4ec-kube-api-access-4676c" (OuterVolumeSpecName: "kube-api-access-4676c") pod "0b623f0c-385a-465d-9132-9406dfdec4ec" (UID: "0b623f0c-385a-465d-9132-9406dfdec4ec"). InnerVolumeSpecName "kube-api-access-4676c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.813935 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.827141 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.827356 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.843573 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8wqh7"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.850457 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0b623f0c-385a-465d-9132-9406dfdec4ec" (UID: "0b623f0c-385a-465d-9132-9406dfdec4ec"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.856133 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.865682 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-trpzx"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.867204 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.873909 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.874043 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-vd664" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.876933 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-55nwz"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.878699 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.895731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-logs\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.895778 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-config-data\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd7mt\" (UniqueName: \"kubernetes.io/projected/4022c6cb-7074-46e7-a366-7005553fa97e-kube-api-access-pd7mt\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4022c6cb-7074-46e7-a366-7005553fa97e-horizon-secret-key\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-config-data\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-scripts\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-combined-ca-bundle\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4022c6cb-7074-46e7-a366-7005553fa97e-logs\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896730 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-scripts\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xswzr\" (UniqueName: \"kubernetes.io/projected/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-kube-api-access-xswzr\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896793 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4676c\" (UniqueName: \"kubernetes.io/projected/0b623f0c-385a-465d-9132-9406dfdec4ec-kube-api-access-4676c\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.896808 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.901934 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-config-data\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.903842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-scripts\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.904530 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4022c6cb-7074-46e7-a366-7005553fa97e-logs\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.923561 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4022c6cb-7074-46e7-a366-7005553fa97e-horizon-secret-key\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.935628 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd7mt\" (UniqueName: \"kubernetes.io/projected/4022c6cb-7074-46e7-a366-7005553fa97e-kube-api-access-pd7mt\") pod \"horizon-6bccbd8f7f-rjdmz\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.936080 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-95tmx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.944798 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-trpzx"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:35.993220 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b623f0c-385a-465d-9132-9406dfdec4ec" (UID: "0b623f0c-385a-465d-9132-9406dfdec4ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.001126 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-scripts\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.001159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.001205 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7v5z\" (UniqueName: \"kubernetes.io/projected/42c61759-0be2-4ae3-95ad-9c0f761b110a-kube-api-access-c7v5z\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.001225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xswzr\" (UniqueName: \"kubernetes.io/projected/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-kube-api-access-xswzr\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.002121 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.003352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-run-httpd\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.010453 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-scripts\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.011226 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8857bfdc9-phjfh"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.015754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.015940 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.018594 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmccx\" (UniqueName: \"kubernetes.io/projected/0781aef7-bf19-478e-931f-a10a3dcd586d-kube-api-access-dmccx\") pod \"barbican-db-sync-trpzx\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.018676 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-logs\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.018699 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-log-httpd\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.018748 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.018784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-scripts\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.018848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.018878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-config-data\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.018906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-config\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.018931 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.019080 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-combined-ca-bundle\") pod \"barbican-db-sync-trpzx\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.019114 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flfrr\" (UniqueName: \"kubernetes.io/projected/597521ee-8310-4c63-bdd2-aeef878fb8b9-kube-api-access-flfrr\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.019155 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.019166 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-config-data" (OuterVolumeSpecName: "config-data") pod "0b623f0c-385a-465d-9132-9406dfdec4ec" (UID: "0b623f0c-385a-465d-9132-9406dfdec4ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.019190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-config-data\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.019541 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-logs\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.021078 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-combined-ca-bundle\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.024316 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-config-data\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.025863 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-db-sync-config-data\") pod \"barbican-db-sync-trpzx\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.026389 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.026418 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b623f0c-385a-465d-9132-9406dfdec4ec-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.028711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xswzr\" (UniqueName: \"kubernetes.io/projected/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-kube-api-access-xswzr\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.031515 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8857bfdc9-phjfh"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.033361 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-combined-ca-bundle\") pod \"placement-db-sync-8wqh7\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.048643 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-55nwz"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.066748 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8wqh7" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132128 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-combined-ca-bundle\") pod \"barbican-db-sync-trpzx\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flfrr\" (UniqueName: \"kubernetes.io/projected/597521ee-8310-4c63-bdd2-aeef878fb8b9-kube-api-access-flfrr\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132311 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-config-data\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfxll\" (UniqueName: \"kubernetes.io/projected/317fbcea-64fb-4132-b028-4b760b98f8b3-kube-api-access-zfxll\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132461 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-db-sync-config-data\") pod \"barbican-db-sync-trpzx\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132521 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7v5z\" (UniqueName: \"kubernetes.io/projected/42c61759-0be2-4ae3-95ad-9c0f761b110a-kube-api-access-c7v5z\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132579 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-run-httpd\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmccx\" (UniqueName: \"kubernetes.io/projected/0781aef7-bf19-478e-931f-a10a3dcd586d-kube-api-access-dmccx\") pod \"barbican-db-sync-trpzx\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132711 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-log-httpd\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-scripts\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132792 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-config-data\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132845 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132874 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-scripts\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132908 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-config\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.132994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.133019 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/317fbcea-64fb-4132-b028-4b760b98f8b3-horizon-secret-key\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.133763 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/317fbcea-64fb-4132-b028-4b760b98f8b3-logs\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.134325 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-run-httpd\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.135602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-log-httpd\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.136465 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.137428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.137787 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-config\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.139048 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.145470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.149278 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-combined-ca-bundle\") pod \"barbican-db-sync-trpzx\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.150237 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-config-data\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.151469 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.154689 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-scripts\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.155729 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-db-sync-config-data\") pod \"barbican-db-sync-trpzx\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.156055 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.156654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7v5z\" (UniqueName: \"kubernetes.io/projected/42c61759-0be2-4ae3-95ad-9c0f761b110a-kube-api-access-c7v5z\") pod \"dnsmasq-dns-58dd9ff6bc-55nwz\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.157406 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmccx\" (UniqueName: \"kubernetes.io/projected/0781aef7-bf19-478e-931f-a10a3dcd586d-kube-api-access-dmccx\") pod \"barbican-db-sync-trpzx\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.160506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flfrr\" (UniqueName: \"kubernetes.io/projected/597521ee-8310-4c63-bdd2-aeef878fb8b9-kube-api-access-flfrr\") pod \"ceilometer-0\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.222693 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.237069 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-scripts\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.237335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-config-data\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.237376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/317fbcea-64fb-4132-b028-4b760b98f8b3-horizon-secret-key\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.237405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/317fbcea-64fb-4132-b028-4b760b98f8b3-logs\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.237455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfxll\" (UniqueName: \"kubernetes.io/projected/317fbcea-64fb-4132-b028-4b760b98f8b3-kube-api-access-zfxll\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.237817 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-scripts\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.238429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/317fbcea-64fb-4132-b028-4b760b98f8b3-logs\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.238738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-config-data\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.239302 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-trpzx" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.253436 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/317fbcea-64fb-4132-b028-4b760b98f8b3-horizon-secret-key\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.253702 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.277885 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfxll\" (UniqueName: \"kubernetes.io/projected/317fbcea-64fb-4132-b028-4b760b98f8b3-kube-api-access-zfxll\") pod \"horizon-8857bfdc9-phjfh\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.348976 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.773054 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sb5fc" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.785090 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-dck68" podUID="ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" containerName="dnsmasq-dns" containerID="cri-o://1b77ef41c7a8f20030e509eecf73388ea0af3c3a34013d5c18bcfae2258ce4f9" gracePeriod=10 Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.787022 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dck68" event={"ID":"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4","Type":"ContainerStarted","Data":"1b77ef41c7a8f20030e509eecf73388ea0af3c3a34013d5c18bcfae2258ce4f9"} Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.787807 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.842731 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-dck68" podStartSLOduration=4.842712338 podStartE2EDuration="4.842712338s" podCreationTimestamp="2025-11-24 13:33:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:36.836836737 +0000 UTC m=+1158.476376047" watchObservedRunningTime="2025-11-24 13:33:36.842712338 +0000 UTC m=+1158.482251648" Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.924593 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-69zcz"] Nov 24 13:33:36 crc kubenswrapper[4824]: I1124 13:33:36.955527 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-95tmx"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.122298 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8wqh7"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.125929 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r2442"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.131868 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bccbd8f7f-rjdmz"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.147340 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-85frw"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.306589 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-55nwz"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.429872 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-ktmj4"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.431302 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.482316 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-ktmj4"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.508691 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-config\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.508747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.508772 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh6dp\" (UniqueName: \"kubernetes.io/projected/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-kube-api-access-dh6dp\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.508822 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.508872 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.508930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.522962 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.563916 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8857bfdc9-phjfh"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.616003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.616279 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-config\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.616302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.616334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh6dp\" (UniqueName: \"kubernetes.io/projected/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-kube-api-access-dh6dp\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.616373 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.616432 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.617710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.618592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.620435 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.620771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.621586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-config\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.694785 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh6dp\" (UniqueName: \"kubernetes.io/projected/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-kube-api-access-dh6dp\") pod \"dnsmasq-dns-785d8bcb8c-ktmj4\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.816324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.822177 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8857bfdc9-phjfh" event={"ID":"317fbcea-64fb-4132-b028-4b760b98f8b3","Type":"ContainerStarted","Data":"b82e79ad288629d9b6cd95ca2b66618a6ba21871b6feb7b9cb424f85f9911bd0"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.830535 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8wqh7" event={"ID":"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce","Type":"ContainerStarted","Data":"cec8ea16c244afd03a966f0e00c5a5a7d272e17cafe0555835df2ea30d2ba54e"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.875061 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-trpzx"] Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.879792 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-95tmx" event={"ID":"8e88a010-ca80-44fc-bd45-d741e3acdc47","Type":"ContainerStarted","Data":"4fe10b3cae19503781c7eea75a17365051a790c7b797f76de1ca4c3a7db834a3"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.879850 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-95tmx" event={"ID":"8e88a010-ca80-44fc-bd45-d741e3acdc47","Type":"ContainerStarted","Data":"05ac447074bc5fa703ab08b1b9118885c9bc27fbd76f38087b7a05ae5bb3430f"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.912056 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-69zcz" event={"ID":"b1af3874-3566-4785-a5fe-13375aa658c3","Type":"ContainerStarted","Data":"38d5c778c8d8e21b4fcfffc530c3cd1bad8dd322e0cba7ad045f473dc3875cff"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.929721 4824 generic.go:334] "Generic (PLEG): container finished" podID="ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" containerID="1b77ef41c7a8f20030e509eecf73388ea0af3c3a34013d5c18bcfae2258ce4f9" exitCode=0 Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.929782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dck68" event={"ID":"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4","Type":"ContainerDied","Data":"1b77ef41c7a8f20030e509eecf73388ea0af3c3a34013d5c18bcfae2258ce4f9"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.932871 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bccbd8f7f-rjdmz" event={"ID":"4022c6cb-7074-46e7-a366-7005553fa97e","Type":"ContainerStarted","Data":"9f6827e17ee5e398d973d7f21f79dbe6a1d252ba2ca7ac4d6060514fc512700c"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.940766 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"597521ee-8310-4c63-bdd2-aeef878fb8b9","Type":"ContainerStarted","Data":"39a54b897d88ef1c661450f89c8643edd5eaed6374f68d36b83b648c539e977b"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.959155 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-85frw" event={"ID":"a4fa0482-6635-4253-b922-5251c70c79d7","Type":"ContainerStarted","Data":"8c5d1c0ce01c1a4695f9e50c984e64ddad91199a371c281218dc461862151413"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.981956 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r2442" event={"ID":"ae1d0586-93a9-4b78-8d56-9826b90e8f25","Type":"ContainerStarted","Data":"c8a1a3455e6f6c0c3b5919f99427cbab96f94a517d88a097709de0ae63e2e964"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.982005 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r2442" event={"ID":"ae1d0586-93a9-4b78-8d56-9826b90e8f25","Type":"ContainerStarted","Data":"3ab87c59089ac1f53ae8fc041e11256b0618779ae53e1bc57c759dadb238e3b9"} Nov 24 13:33:37 crc kubenswrapper[4824]: I1124 13:33:37.992094 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.006030 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-r2442" podStartSLOduration=4.006014374 podStartE2EDuration="4.006014374s" podCreationTimestamp="2025-11-24 13:33:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:38.004201387 +0000 UTC m=+1159.643740697" watchObservedRunningTime="2025-11-24 13:33:38.006014374 +0000 UTC m=+1159.645553684" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.012184 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-95tmx" podStartSLOduration=3.012162612 podStartE2EDuration="3.012162612s" podCreationTimestamp="2025-11-24 13:33:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:37.912315181 +0000 UTC m=+1159.551854491" watchObservedRunningTime="2025-11-24 13:33:38.012162612 +0000 UTC m=+1159.651701922" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.022369 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-55nwz"] Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.124449 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-sb\") pod \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.124763 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-config\") pod \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.124793 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r97nc\" (UniqueName: \"kubernetes.io/projected/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-kube-api-access-r97nc\") pod \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.124898 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-swift-storage-0\") pod \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.124924 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-nb\") pod \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.124963 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-svc\") pod \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\" (UID: \"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4\") " Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.152349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-kube-api-access-r97nc" (OuterVolumeSpecName: "kube-api-access-r97nc") pod "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" (UID: "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4"). InnerVolumeSpecName "kube-api-access-r97nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.226943 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r97nc\" (UniqueName: \"kubernetes.io/projected/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-kube-api-access-r97nc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.234319 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" (UID: "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.238547 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" (UID: "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.266433 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-config" (OuterVolumeSpecName: "config") pod "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" (UID: "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.324997 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" (UID: "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.331247 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.331275 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.331283 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.331291 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.367795 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:38 crc kubenswrapper[4824]: E1124 13:33:38.368146 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" containerName="dnsmasq-dns" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.368156 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" containerName="dnsmasq-dns" Nov 24 13:33:38 crc kubenswrapper[4824]: E1124 13:33:38.368175 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" containerName="init" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.368181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" containerName="init" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.368352 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" containerName="dnsmasq-dns" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.369157 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.373322 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" (UID: "ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.381741 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.382833 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-p7tpr" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.392816 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.415212 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.433618 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.433692 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.433742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.433800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-logs\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.433867 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.433901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwmj8\" (UniqueName: \"kubernetes.io/projected/3303f1e8-4944-430e-93e6-31bafe0e8c6b-kube-api-access-wwmj8\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.433937 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.433996 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.535895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwmj8\" (UniqueName: \"kubernetes.io/projected/3303f1e8-4944-430e-93e6-31bafe0e8c6b-kube-api-access-wwmj8\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.536167 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.536209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.536229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.536270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.536304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-logs\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.536356 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.543675 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.544099 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.550481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.550546 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-ktmj4"] Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.550850 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-logs\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.555593 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.574650 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.594895 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.596591 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.601172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwmj8\" (UniqueName: \"kubernetes.io/projected/3303f1e8-4944-430e-93e6-31bafe0e8c6b-kube-api-access-wwmj8\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.601978 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.626016 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.638790 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.742874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.742973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qz2m\" (UniqueName: \"kubernetes.io/projected/affd78e4-4469-4484-80dd-598dfb8ff929-kube-api-access-9qz2m\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.742999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-scripts\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.743057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-logs\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.743083 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.743101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-config-data\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.743155 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.762930 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.763714 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.792759 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6bccbd8f7f-rjdmz"] Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.846064 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-logs\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.846329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.846350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-config-data\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.846391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.846465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.846538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qz2m\" (UniqueName: \"kubernetes.io/projected/affd78e4-4469-4484-80dd-598dfb8ff929-kube-api-access-9qz2m\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.846560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-scripts\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.857206 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-logs\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.860882 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.860991 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.881455 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-scripts\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.881787 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-config-data\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.913633 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-558844779c-bw4xd"] Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.915189 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.915744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.947900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/240e0890-6e9c-4b0b-95ce-786769479b06-logs\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.948104 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-config-data\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.948216 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-scripts\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.948282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/240e0890-6e9c-4b0b-95ce-786769479b06-horizon-secret-key\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.948379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6lkd\" (UniqueName: \"kubernetes.io/projected/240e0890-6e9c-4b0b-95ce-786769479b06-kube-api-access-p6lkd\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:38 crc kubenswrapper[4824]: I1124 13:33:38.976187 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-558844779c-bw4xd"] Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.002901 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.019955 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qz2m\" (UniqueName: \"kubernetes.io/projected/affd78e4-4469-4484-80dd-598dfb8ff929-kube-api-access-9qz2m\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:39 crc kubenswrapper[4824]: E1124 13:33:39.064630 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[glance kube-api-access-9qz2m], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="affd78e4-4469-4484-80dd-598dfb8ff929" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.065570 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/240e0890-6e9c-4b0b-95ce-786769479b06-logs\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.065786 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-config-data\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.066025 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-scripts\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.066097 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/240e0890-6e9c-4b0b-95ce-786769479b06-horizon-secret-key\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.066216 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6lkd\" (UniqueName: \"kubernetes.io/projected/240e0890-6e9c-4b0b-95ce-786769479b06-kube-api-access-p6lkd\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.067311 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/240e0890-6e9c-4b0b-95ce-786769479b06-logs\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.069542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-scripts\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.070953 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.074341 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-config-data\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.100784 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.101490 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-dck68" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.110744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" event={"ID":"42c61759-0be2-4ae3-95ad-9c0f761b110a","Type":"ContainerStarted","Data":"993caf657368fb1182d803a4e69e97d466b31e734286bc1fa1f180dfb65d986d"} Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.110867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dck68" event={"ID":"ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4","Type":"ContainerDied","Data":"ca2dc56148a435952d950c6d93e4ce558c6a52b62de414c4aca8611e84385158"} Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.110892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-trpzx" event={"ID":"0781aef7-bf19-478e-931f-a10a3dcd586d","Type":"ContainerStarted","Data":"8e7202915d7cd0d11d7d5f6752bf8cf623f1c2956ad82d39fd28e623d2643aca"} Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.110924 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" event={"ID":"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc","Type":"ContainerStarted","Data":"bd0d8b4c1719af343a39a7ecb0d7947eff371b58bd181bd3c09e54b3350ad8f7"} Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.110936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-85frw" event={"ID":"a4fa0482-6635-4253-b922-5251c70c79d7","Type":"ContainerDied","Data":"8ee938e091d32ed81afcd00483d952348faa2b5d5cc92000826b9bd2884484a8"} Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.110956 4824 scope.go:117] "RemoveContainer" containerID="1b77ef41c7a8f20030e509eecf73388ea0af3c3a34013d5c18bcfae2258ce4f9" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.106878 4824 generic.go:334] "Generic (PLEG): container finished" podID="a4fa0482-6635-4253-b922-5251c70c79d7" containerID="8ee938e091d32ed81afcd00483d952348faa2b5d5cc92000826b9bd2884484a8" exitCode=0 Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.105491 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6lkd\" (UniqueName: \"kubernetes.io/projected/240e0890-6e9c-4b0b-95ce-786769479b06-kube-api-access-p6lkd\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.121404 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/240e0890-6e9c-4b0b-95ce-786769479b06-horizon-secret-key\") pod \"horizon-558844779c-bw4xd\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.247467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.290040 4824 scope.go:117] "RemoveContainer" containerID="86b0dd4a57a4f860fb9d2ff2df5a29b0ec615e8bb84358debd2211cd17c64c48" Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.712697 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dck68"] Nov 24 13:33:39 crc kubenswrapper[4824]: I1124 13:33:39.735889 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dck68"] Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.098417 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.155945 4824 generic.go:334] "Generic (PLEG): container finished" podID="42c61759-0be2-4ae3-95ad-9c0f761b110a" containerID="fd6d84bf0f20ea176ec11397db6d55b0d400c85fffca7338b401340982144d19" exitCode=0 Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.156214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" event={"ID":"42c61759-0be2-4ae3-95ad-9c0f761b110a","Type":"ContainerDied","Data":"fd6d84bf0f20ea176ec11397db6d55b0d400c85fffca7338b401340982144d19"} Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.187609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.188562 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" event={"ID":"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc","Type":"ContainerStarted","Data":"7012cee1033aa5aa4ff422eeed785e1227e70a31c8fcfda4dd31833432ac72a4"} Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.260113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.269271 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkqf9\" (UniqueName: \"kubernetes.io/projected/a4fa0482-6635-4253-b922-5251c70c79d7-kube-api-access-hkqf9\") pod \"a4fa0482-6635-4253-b922-5251c70c79d7\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325253 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-config\") pod \"a4fa0482-6635-4253-b922-5251c70c79d7\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-logs\") pod \"affd78e4-4469-4484-80dd-598dfb8ff929\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325399 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-httpd-run\") pod \"affd78e4-4469-4484-80dd-598dfb8ff929\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325418 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-svc\") pod \"a4fa0482-6635-4253-b922-5251c70c79d7\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325446 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-combined-ca-bundle\") pod \"affd78e4-4469-4484-80dd-598dfb8ff929\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325485 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qz2m\" (UniqueName: \"kubernetes.io/projected/affd78e4-4469-4484-80dd-598dfb8ff929-kube-api-access-9qz2m\") pod \"affd78e4-4469-4484-80dd-598dfb8ff929\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325509 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-nb\") pod \"a4fa0482-6635-4253-b922-5251c70c79d7\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325525 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-swift-storage-0\") pod \"a4fa0482-6635-4253-b922-5251c70c79d7\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325539 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-sb\") pod \"a4fa0482-6635-4253-b922-5251c70c79d7\" (UID: \"a4fa0482-6635-4253-b922-5251c70c79d7\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325578 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"affd78e4-4469-4484-80dd-598dfb8ff929\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325596 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-config-data\") pod \"affd78e4-4469-4484-80dd-598dfb8ff929\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.325620 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-scripts\") pod \"affd78e4-4469-4484-80dd-598dfb8ff929\" (UID: \"affd78e4-4469-4484-80dd-598dfb8ff929\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.329944 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "affd78e4-4469-4484-80dd-598dfb8ff929" (UID: "affd78e4-4469-4484-80dd-598dfb8ff929"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.330232 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-logs" (OuterVolumeSpecName: "logs") pod "affd78e4-4469-4484-80dd-598dfb8ff929" (UID: "affd78e4-4469-4484-80dd-598dfb8ff929"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.368892 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4fa0482-6635-4253-b922-5251c70c79d7-kube-api-access-hkqf9" (OuterVolumeSpecName: "kube-api-access-hkqf9") pod "a4fa0482-6635-4253-b922-5251c70c79d7" (UID: "a4fa0482-6635-4253-b922-5251c70c79d7"). InnerVolumeSpecName "kube-api-access-hkqf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.371548 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "affd78e4-4469-4484-80dd-598dfb8ff929" (UID: "affd78e4-4469-4484-80dd-598dfb8ff929"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.377441 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-scripts" (OuterVolumeSpecName: "scripts") pod "affd78e4-4469-4484-80dd-598dfb8ff929" (UID: "affd78e4-4469-4484-80dd-598dfb8ff929"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.377728 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "affd78e4-4469-4484-80dd-598dfb8ff929" (UID: "affd78e4-4469-4484-80dd-598dfb8ff929"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.395246 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-558844779c-bw4xd"] Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.406444 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-config-data" (OuterVolumeSpecName: "config-data") pod "affd78e4-4469-4484-80dd-598dfb8ff929" (UID: "affd78e4-4469-4484-80dd-598dfb8ff929"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.415035 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/affd78e4-4469-4484-80dd-598dfb8ff929-kube-api-access-9qz2m" (OuterVolumeSpecName: "kube-api-access-9qz2m") pod "affd78e4-4469-4484-80dd-598dfb8ff929" (UID: "affd78e4-4469-4484-80dd-598dfb8ff929"). InnerVolumeSpecName "kube-api-access-9qz2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.448567 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.448596 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qz2m\" (UniqueName: \"kubernetes.io/projected/affd78e4-4469-4484-80dd-598dfb8ff929-kube-api-access-9qz2m\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.448618 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.448629 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.448638 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affd78e4-4469-4484-80dd-598dfb8ff929-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.448647 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkqf9\" (UniqueName: \"kubernetes.io/projected/a4fa0482-6635-4253-b922-5251c70c79d7-kube-api-access-hkqf9\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.448655 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.448666 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/affd78e4-4469-4484-80dd-598dfb8ff929-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.522218 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-config" (OuterVolumeSpecName: "config") pod "a4fa0482-6635-4253-b922-5251c70c79d7" (UID: "a4fa0482-6635-4253-b922-5251c70c79d7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.546040 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a4fa0482-6635-4253-b922-5251c70c79d7" (UID: "a4fa0482-6635-4253-b922-5251c70c79d7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.550787 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.550831 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.571971 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a4fa0482-6635-4253-b922-5251c70c79d7" (UID: "a4fa0482-6635-4253-b922-5251c70c79d7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.572543 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a4fa0482-6635-4253-b922-5251c70c79d7" (UID: "a4fa0482-6635-4253-b922-5251c70c79d7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.580402 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a4fa0482-6635-4253-b922-5251c70c79d7" (UID: "a4fa0482-6635-4253-b922-5251c70c79d7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.607251 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.662739 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.662772 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.662786 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4fa0482-6635-4253-b922-5251c70c79d7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.662798 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.764363 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.870585 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-sb\") pod \"42c61759-0be2-4ae3-95ad-9c0f761b110a\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.870708 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7v5z\" (UniqueName: \"kubernetes.io/projected/42c61759-0be2-4ae3-95ad-9c0f761b110a-kube-api-access-c7v5z\") pod \"42c61759-0be2-4ae3-95ad-9c0f761b110a\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.870757 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-svc\") pod \"42c61759-0be2-4ae3-95ad-9c0f761b110a\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.870788 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-nb\") pod \"42c61759-0be2-4ae3-95ad-9c0f761b110a\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.870873 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-swift-storage-0\") pod \"42c61759-0be2-4ae3-95ad-9c0f761b110a\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.870972 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-config\") pod \"42c61759-0be2-4ae3-95ad-9c0f761b110a\" (UID: \"42c61759-0be2-4ae3-95ad-9c0f761b110a\") " Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.915434 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42c61759-0be2-4ae3-95ad-9c0f761b110a-kube-api-access-c7v5z" (OuterVolumeSpecName: "kube-api-access-c7v5z") pod "42c61759-0be2-4ae3-95ad-9c0f761b110a" (UID: "42c61759-0be2-4ae3-95ad-9c0f761b110a"). InnerVolumeSpecName "kube-api-access-c7v5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.974015 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7v5z\" (UniqueName: \"kubernetes.io/projected/42c61759-0be2-4ae3-95ad-9c0f761b110a-kube-api-access-c7v5z\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:40 crc kubenswrapper[4824]: I1124 13:33:40.982382 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "42c61759-0be2-4ae3-95ad-9c0f761b110a" (UID: "42c61759-0be2-4ae3-95ad-9c0f761b110a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.023168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "42c61759-0be2-4ae3-95ad-9c0f761b110a" (UID: "42c61759-0be2-4ae3-95ad-9c0f761b110a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.035244 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "42c61759-0be2-4ae3-95ad-9c0f761b110a" (UID: "42c61759-0be2-4ae3-95ad-9c0f761b110a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.046023 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4" path="/var/lib/kubelet/pods/ae2bc45a-54d4-4d07-8c1d-5c48cc9000b4/volumes" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.054090 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-config" (OuterVolumeSpecName: "config") pod "42c61759-0be2-4ae3-95ad-9c0f761b110a" (UID: "42c61759-0be2-4ae3-95ad-9c0f761b110a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.059488 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "42c61759-0be2-4ae3-95ad-9c0f761b110a" (UID: "42c61759-0be2-4ae3-95ad-9c0f761b110a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.075860 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.077190 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.077202 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.077215 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.077223 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42c61759-0be2-4ae3-95ad-9c0f761b110a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.252064 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3303f1e8-4944-430e-93e6-31bafe0e8c6b","Type":"ContainerStarted","Data":"9b7537adc8776d2c889ac3ff94af26c9c0ac809ea5b1170598f6a9092caa8855"} Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.259914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" event={"ID":"42c61759-0be2-4ae3-95ad-9c0f761b110a","Type":"ContainerDied","Data":"993caf657368fb1182d803a4e69e97d466b31e734286bc1fa1f180dfb65d986d"} Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.259957 4824 scope.go:117] "RemoveContainer" containerID="fd6d84bf0f20ea176ec11397db6d55b0d400c85fffca7338b401340982144d19" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.260059 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-55nwz" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.290004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-558844779c-bw4xd" event={"ID":"240e0890-6e9c-4b0b-95ce-786769479b06","Type":"ContainerStarted","Data":"b4e357bf17f6920b3b15a07765e563542f08c534c433ab0e7f51da9475a79e5e"} Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.300419 4824 generic.go:334] "Generic (PLEG): container finished" podID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" containerID="7012cee1033aa5aa4ff422eeed785e1227e70a31c8fcfda4dd31833432ac72a4" exitCode=0 Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.300596 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" event={"ID":"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc","Type":"ContainerDied","Data":"7012cee1033aa5aa4ff422eeed785e1227e70a31c8fcfda4dd31833432ac72a4"} Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.300655 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" event={"ID":"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc","Type":"ContainerStarted","Data":"8a39ea221e0a460c01cbe5046e500c3a42d4d725d4b7814483f2c840432779c8"} Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.301148 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.316656 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.317411 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-85frw" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.317749 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-85frw" event={"ID":"a4fa0482-6635-4253-b922-5251c70c79d7","Type":"ContainerDied","Data":"8c5d1c0ce01c1a4695f9e50c984e64ddad91199a371c281218dc461862151413"} Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.360639 4824 scope.go:117] "RemoveContainer" containerID="8ee938e091d32ed81afcd00483d952348faa2b5d5cc92000826b9bd2884484a8" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.361968 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-55nwz"] Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.371433 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-55nwz"] Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.424208 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" podStartSLOduration=4.424185584 podStartE2EDuration="4.424185584s" podCreationTimestamp="2025-11-24 13:33:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:41.409118666 +0000 UTC m=+1163.048657976" watchObservedRunningTime="2025-11-24 13:33:41.424185584 +0000 UTC m=+1163.063724904" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.457555 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.464318 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.519462 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:41 crc kubenswrapper[4824]: E1124 13:33:41.519850 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c61759-0be2-4ae3-95ad-9c0f761b110a" containerName="init" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.519861 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c61759-0be2-4ae3-95ad-9c0f761b110a" containerName="init" Nov 24 13:33:41 crc kubenswrapper[4824]: E1124 13:33:41.519913 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4fa0482-6635-4253-b922-5251c70c79d7" containerName="init" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.519919 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4fa0482-6635-4253-b922-5251c70c79d7" containerName="init" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.527786 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42c61759-0be2-4ae3-95ad-9c0f761b110a" containerName="init" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.527840 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4fa0482-6635-4253-b922-5251c70c79d7" containerName="init" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.528969 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.543251 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.545242 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-85frw"] Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.561566 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-85frw"] Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.585855 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.696102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.696155 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.696184 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69q58\" (UniqueName: \"kubernetes.io/projected/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-kube-api-access-69q58\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.696234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.696284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-logs\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.696315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.696441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.797655 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-logs\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.797711 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.797745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.797778 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.797802 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.797881 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69q58\" (UniqueName: \"kubernetes.io/projected/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-kube-api-access-69q58\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.797923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.798155 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-logs\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.798181 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.800303 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.814397 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.829180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.840252 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.899331 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:41 crc kubenswrapper[4824]: I1124 13:33:41.919579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69q58\" (UniqueName: \"kubernetes.io/projected/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-kube-api-access-69q58\") pod \"glance-default-internal-api-0\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:33:42 crc kubenswrapper[4824]: I1124 13:33:42.166225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:33:42 crc kubenswrapper[4824]: I1124 13:33:42.345785 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3303f1e8-4944-430e-93e6-31bafe0e8c6b","Type":"ContainerStarted","Data":"0867909c9e672d429b921035060badefdb63e7e95df94652cbbe24560f991555"} Nov 24 13:33:42 crc kubenswrapper[4824]: I1124 13:33:42.956694 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:43 crc kubenswrapper[4824]: I1124 13:33:43.021995 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42c61759-0be2-4ae3-95ad-9c0f761b110a" path="/var/lib/kubelet/pods/42c61759-0be2-4ae3-95ad-9c0f761b110a/volumes" Nov 24 13:33:43 crc kubenswrapper[4824]: I1124 13:33:43.022855 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4fa0482-6635-4253-b922-5251c70c79d7" path="/var/lib/kubelet/pods/a4fa0482-6635-4253-b922-5251c70c79d7/volumes" Nov 24 13:33:43 crc kubenswrapper[4824]: I1124 13:33:43.024714 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="affd78e4-4469-4484-80dd-598dfb8ff929" path="/var/lib/kubelet/pods/affd78e4-4469-4484-80dd-598dfb8ff929/volumes" Nov 24 13:33:43 crc kubenswrapper[4824]: I1124 13:33:43.379181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe","Type":"ContainerStarted","Data":"26f78abc9b893dd7339573f4b49fbc7ba563f6d29d63b4ea9eb251e2e0c2032c"} Nov 24 13:33:44 crc kubenswrapper[4824]: I1124 13:33:44.421922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3303f1e8-4944-430e-93e6-31bafe0e8c6b","Type":"ContainerStarted","Data":"a94b804e071c1782ef5b773b77d9f36a285e09bd557397e071100fbe6777b31d"} Nov 24 13:33:44 crc kubenswrapper[4824]: I1124 13:33:44.422029 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerName="glance-log" containerID="cri-o://0867909c9e672d429b921035060badefdb63e7e95df94652cbbe24560f991555" gracePeriod=30 Nov 24 13:33:44 crc kubenswrapper[4824]: I1124 13:33:44.422063 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerName="glance-httpd" containerID="cri-o://a94b804e071c1782ef5b773b77d9f36a285e09bd557397e071100fbe6777b31d" gracePeriod=30 Nov 24 13:33:44 crc kubenswrapper[4824]: I1124 13:33:44.436007 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe","Type":"ContainerStarted","Data":"b629cf93ec1fb2d48832ab8317c67ddf23fe380fbec6a5784ef5de4711a8d167"} Nov 24 13:33:44 crc kubenswrapper[4824]: I1124 13:33:44.451643 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.451620622 podStartE2EDuration="7.451620622s" podCreationTimestamp="2025-11-24 13:33:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:44.447935467 +0000 UTC m=+1166.087474807" watchObservedRunningTime="2025-11-24 13:33:44.451620622 +0000 UTC m=+1166.091159942" Nov 24 13:33:45 crc kubenswrapper[4824]: I1124 13:33:45.451183 4824 generic.go:334] "Generic (PLEG): container finished" podID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerID="a94b804e071c1782ef5b773b77d9f36a285e09bd557397e071100fbe6777b31d" exitCode=0 Nov 24 13:33:45 crc kubenswrapper[4824]: I1124 13:33:45.451220 4824 generic.go:334] "Generic (PLEG): container finished" podID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerID="0867909c9e672d429b921035060badefdb63e7e95df94652cbbe24560f991555" exitCode=143 Nov 24 13:33:45 crc kubenswrapper[4824]: I1124 13:33:45.451246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3303f1e8-4944-430e-93e6-31bafe0e8c6b","Type":"ContainerDied","Data":"a94b804e071c1782ef5b773b77d9f36a285e09bd557397e071100fbe6777b31d"} Nov 24 13:33:45 crc kubenswrapper[4824]: I1124 13:33:45.451312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3303f1e8-4944-430e-93e6-31bafe0e8c6b","Type":"ContainerDied","Data":"0867909c9e672d429b921035060badefdb63e7e95df94652cbbe24560f991555"} Nov 24 13:33:45 crc kubenswrapper[4824]: I1124 13:33:45.458008 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe","Type":"ContainerStarted","Data":"fc8974a7cf2e902af051b1559277aecae1a57a482335156560c839bcd1a2d5f6"} Nov 24 13:33:45 crc kubenswrapper[4824]: I1124 13:33:45.475166 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.475150887 podStartE2EDuration="4.475150887s" podCreationTimestamp="2025-11-24 13:33:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:33:45.47371052 +0000 UTC m=+1167.113249830" watchObservedRunningTime="2025-11-24 13:33:45.475150887 +0000 UTC m=+1167.114690197" Nov 24 13:33:46 crc kubenswrapper[4824]: I1124 13:33:46.468061 4824 generic.go:334] "Generic (PLEG): container finished" podID="ae1d0586-93a9-4b78-8d56-9826b90e8f25" containerID="c8a1a3455e6f6c0c3b5919f99427cbab96f94a517d88a097709de0ae63e2e964" exitCode=0 Nov 24 13:33:46 crc kubenswrapper[4824]: I1124 13:33:46.468169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r2442" event={"ID":"ae1d0586-93a9-4b78-8d56-9826b90e8f25","Type":"ContainerDied","Data":"c8a1a3455e6f6c0c3b5919f99427cbab96f94a517d88a097709de0ae63e2e964"} Nov 24 13:33:47 crc kubenswrapper[4824]: I1124 13:33:47.144589 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:33:47 crc kubenswrapper[4824]: I1124 13:33:47.477702 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerName="glance-log" containerID="cri-o://b629cf93ec1fb2d48832ab8317c67ddf23fe380fbec6a5784ef5de4711a8d167" gracePeriod=30 Nov 24 13:33:47 crc kubenswrapper[4824]: I1124 13:33:47.480705 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerName="glance-httpd" containerID="cri-o://fc8974a7cf2e902af051b1559277aecae1a57a482335156560c839bcd1a2d5f6" gracePeriod=30 Nov 24 13:33:47 crc kubenswrapper[4824]: I1124 13:33:47.818014 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:33:47 crc kubenswrapper[4824]: I1124 13:33:47.871493 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-74w28"] Nov 24 13:33:47 crc kubenswrapper[4824]: I1124 13:33:47.871960 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-74w28" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" containerID="cri-o://0eacd6c6d75a66c6c694c6209d042d8d8cefc528e587d8758df40d15b776d18d" gracePeriod=10 Nov 24 13:33:48 crc kubenswrapper[4824]: I1124 13:33:48.836106 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-74w28" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: connect: connection refused" Nov 24 13:33:48 crc kubenswrapper[4824]: I1124 13:33:48.874654 4824 generic.go:334] "Generic (PLEG): container finished" podID="429c7611-6a69-43a2-8047-d46b48f48447" containerID="0eacd6c6d75a66c6c694c6209d042d8d8cefc528e587d8758df40d15b776d18d" exitCode=0 Nov 24 13:33:48 crc kubenswrapper[4824]: I1124 13:33:48.874727 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-74w28" event={"ID":"429c7611-6a69-43a2-8047-d46b48f48447","Type":"ContainerDied","Data":"0eacd6c6d75a66c6c694c6209d042d8d8cefc528e587d8758df40d15b776d18d"} Nov 24 13:33:48 crc kubenswrapper[4824]: I1124 13:33:48.878111 4824 generic.go:334] "Generic (PLEG): container finished" podID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerID="fc8974a7cf2e902af051b1559277aecae1a57a482335156560c839bcd1a2d5f6" exitCode=0 Nov 24 13:33:48 crc kubenswrapper[4824]: I1124 13:33:48.878139 4824 generic.go:334] "Generic (PLEG): container finished" podID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerID="b629cf93ec1fb2d48832ab8317c67ddf23fe380fbec6a5784ef5de4711a8d167" exitCode=143 Nov 24 13:33:48 crc kubenswrapper[4824]: I1124 13:33:48.878159 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe","Type":"ContainerDied","Data":"fc8974a7cf2e902af051b1559277aecae1a57a482335156560c839bcd1a2d5f6"} Nov 24 13:33:48 crc kubenswrapper[4824]: I1124 13:33:48.878183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe","Type":"ContainerDied","Data":"b629cf93ec1fb2d48832ab8317c67ddf23fe380fbec6a5784ef5de4711a8d167"} Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.115884 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8857bfdc9-phjfh"] Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.115946 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c96f4646d-wsbgv"] Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.129905 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.135292 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.135903 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-config-data\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.135951 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-combined-ca-bundle\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.136035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-scripts\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.136114 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-tls-certs\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.136136 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-secret-key\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.136153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqp5f\" (UniqueName: \"kubernetes.io/projected/bc2d2fae-b656-464e-b1a8-89b4c5160792-kube-api-access-mqp5f\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.136179 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d2fae-b656-464e-b1a8-89b4c5160792-logs\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.140793 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c96f4646d-wsbgv"] Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.217232 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-558844779c-bw4xd"] Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.238522 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-959bb6dd8-8pcrj"] Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.240213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-config-data\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.240249 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.240255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-combined-ca-bundle\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.240315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-scripts\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.240380 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-tls-certs\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.240399 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-secret-key\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.240420 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqp5f\" (UniqueName: \"kubernetes.io/projected/bc2d2fae-b656-464e-b1a8-89b4c5160792-kube-api-access-mqp5f\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.240449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d2fae-b656-464e-b1a8-89b4c5160792-logs\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.240843 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d2fae-b656-464e-b1a8-89b4c5160792-logs\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.242730 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-config-data\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.245564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-scripts\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.250381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-tls-certs\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.264509 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-combined-ca-bundle\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.275662 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-959bb6dd8-8pcrj"] Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.288624 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-secret-key\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.309539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqp5f\" (UniqueName: \"kubernetes.io/projected/bc2d2fae-b656-464e-b1a8-89b4c5160792-kube-api-access-mqp5f\") pod \"horizon-7c96f4646d-wsbgv\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.342611 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5828a51f-d146-44ea-b8ea-423e017939d8-scripts\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.342705 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qmkw\" (UniqueName: \"kubernetes.io/projected/5828a51f-d146-44ea-b8ea-423e017939d8-kube-api-access-7qmkw\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.342730 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5828a51f-d146-44ea-b8ea-423e017939d8-logs\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.342749 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5828a51f-d146-44ea-b8ea-423e017939d8-combined-ca-bundle\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.342771 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5828a51f-d146-44ea-b8ea-423e017939d8-horizon-secret-key\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.342797 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5828a51f-d146-44ea-b8ea-423e017939d8-config-data\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.342873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5828a51f-d146-44ea-b8ea-423e017939d8-horizon-tls-certs\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.444826 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qmkw\" (UniqueName: \"kubernetes.io/projected/5828a51f-d146-44ea-b8ea-423e017939d8-kube-api-access-7qmkw\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.444876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5828a51f-d146-44ea-b8ea-423e017939d8-logs\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.444904 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5828a51f-d146-44ea-b8ea-423e017939d8-combined-ca-bundle\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.444934 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5828a51f-d146-44ea-b8ea-423e017939d8-horizon-secret-key\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.444966 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5828a51f-d146-44ea-b8ea-423e017939d8-config-data\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.444996 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5828a51f-d146-44ea-b8ea-423e017939d8-horizon-tls-certs\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.445050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5828a51f-d146-44ea-b8ea-423e017939d8-scripts\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.446481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5828a51f-d146-44ea-b8ea-423e017939d8-scripts\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.446709 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5828a51f-d146-44ea-b8ea-423e017939d8-logs\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.448257 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5828a51f-d146-44ea-b8ea-423e017939d8-config-data\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.452204 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5828a51f-d146-44ea-b8ea-423e017939d8-horizon-tls-certs\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.452827 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5828a51f-d146-44ea-b8ea-423e017939d8-horizon-secret-key\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.462418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5828a51f-d146-44ea-b8ea-423e017939d8-combined-ca-bundle\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.465025 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qmkw\" (UniqueName: \"kubernetes.io/projected/5828a51f-d146-44ea-b8ea-423e017939d8-kube-api-access-7qmkw\") pod \"horizon-959bb6dd8-8pcrj\" (UID: \"5828a51f-d146-44ea-b8ea-423e017939d8\") " pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.465410 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:33:49 crc kubenswrapper[4824]: I1124 13:33:49.670062 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.048686 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.224726 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-config-data\") pod \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.224758 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-combined-ca-bundle\") pod \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.224855 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-scripts\") pod \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.224874 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbtcg\" (UniqueName: \"kubernetes.io/projected/ae1d0586-93a9-4b78-8d56-9826b90e8f25-kube-api-access-rbtcg\") pod \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.224903 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-fernet-keys\") pod \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.224963 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-credential-keys\") pod \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\" (UID: \"ae1d0586-93a9-4b78-8d56-9826b90e8f25\") " Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.230619 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ae1d0586-93a9-4b78-8d56-9826b90e8f25" (UID: "ae1d0586-93a9-4b78-8d56-9826b90e8f25"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.231761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae1d0586-93a9-4b78-8d56-9826b90e8f25-kube-api-access-rbtcg" (OuterVolumeSpecName: "kube-api-access-rbtcg") pod "ae1d0586-93a9-4b78-8d56-9826b90e8f25" (UID: "ae1d0586-93a9-4b78-8d56-9826b90e8f25"). InnerVolumeSpecName "kube-api-access-rbtcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.233460 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ae1d0586-93a9-4b78-8d56-9826b90e8f25" (UID: "ae1d0586-93a9-4b78-8d56-9826b90e8f25"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.234941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-scripts" (OuterVolumeSpecName: "scripts") pod "ae1d0586-93a9-4b78-8d56-9826b90e8f25" (UID: "ae1d0586-93a9-4b78-8d56-9826b90e8f25"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.262543 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-config-data" (OuterVolumeSpecName: "config-data") pod "ae1d0586-93a9-4b78-8d56-9826b90e8f25" (UID: "ae1d0586-93a9-4b78-8d56-9826b90e8f25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.307932 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae1d0586-93a9-4b78-8d56-9826b90e8f25" (UID: "ae1d0586-93a9-4b78-8d56-9826b90e8f25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.333614 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.333653 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbtcg\" (UniqueName: \"kubernetes.io/projected/ae1d0586-93a9-4b78-8d56-9826b90e8f25-kube-api-access-rbtcg\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.333668 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.333681 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.333692 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.333702 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1d0586-93a9-4b78-8d56-9826b90e8f25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.936187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r2442" event={"ID":"ae1d0586-93a9-4b78-8d56-9826b90e8f25","Type":"ContainerDied","Data":"3ab87c59089ac1f53ae8fc041e11256b0618779ae53e1bc57c759dadb238e3b9"} Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.936228 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ab87c59089ac1f53ae8fc041e11256b0618779ae53e1bc57c759dadb238e3b9" Nov 24 13:33:53 crc kubenswrapper[4824]: I1124 13:33:53.936287 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r2442" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.127124 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-r2442"] Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.136609 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-r2442"] Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.223121 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nhzcv"] Nov 24 13:33:54 crc kubenswrapper[4824]: E1124 13:33:54.223972 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1d0586-93a9-4b78-8d56-9826b90e8f25" containerName="keystone-bootstrap" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.224655 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1d0586-93a9-4b78-8d56-9826b90e8f25" containerName="keystone-bootstrap" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.225055 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae1d0586-93a9-4b78-8d56-9826b90e8f25" containerName="keystone-bootstrap" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.225890 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.230336 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k9jvt" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.230341 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.230658 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.230744 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.230713 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.235744 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nhzcv"] Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.248390 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-credential-keys\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.248444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-scripts\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.248467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrtpd\" (UniqueName: \"kubernetes.io/projected/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-kube-api-access-mrtpd\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.248554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-combined-ca-bundle\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.248601 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-config-data\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.248676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-fernet-keys\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.350884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-credential-keys\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.350950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-scripts\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.350980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrtpd\" (UniqueName: \"kubernetes.io/projected/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-kube-api-access-mrtpd\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.351025 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-combined-ca-bundle\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.351072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-config-data\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.351154 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-fernet-keys\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.354466 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-credential-keys\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.354566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-scripts\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.356633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-fernet-keys\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.362411 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-combined-ca-bundle\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.363392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-config-data\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.367229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrtpd\" (UniqueName: \"kubernetes.io/projected/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-kube-api-access-mrtpd\") pod \"keystone-bootstrap-nhzcv\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:54 crc kubenswrapper[4824]: I1124 13:33:54.544991 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:33:55 crc kubenswrapper[4824]: I1124 13:33:55.022964 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae1d0586-93a9-4b78-8d56-9826b90e8f25" path="/var/lib/kubelet/pods/ae1d0586-93a9-4b78-8d56-9826b90e8f25/volumes" Nov 24 13:33:58 crc kubenswrapper[4824]: E1124 13:33:58.318961 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 24 13:33:58 crc kubenswrapper[4824]: E1124 13:33:58.319950 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n56fh8hb5hdfh554h66ch6dh4hd6h664h644h69h698h58ch78h667h66ch57dhcfh87h57dh686h8bh576h684hffhbch89h697h587hbbh657q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pd7mt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6bccbd8f7f-rjdmz_openstack(4022c6cb-7074-46e7-a366-7005553fa97e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:33:58 crc kubenswrapper[4824]: E1124 13:33:58.331604 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6bccbd8f7f-rjdmz" podUID="4022c6cb-7074-46e7-a366-7005553fa97e" Nov 24 13:33:58 crc kubenswrapper[4824]: I1124 13:33:58.640539 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-74w28" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: i/o timeout" Nov 24 13:34:01 crc kubenswrapper[4824]: E1124 13:34:01.519769 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 24 13:34:01 crc kubenswrapper[4824]: E1124 13:34:01.520517 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xswzr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-8wqh7_openstack(aef6bc1b-dfa9-4c16-9940-420f9a7a43ce): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:34:01 crc kubenswrapper[4824]: E1124 13:34:01.521794 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-8wqh7" podUID="aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" Nov 24 13:34:01 crc kubenswrapper[4824]: E1124 13:34:01.533354 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 24 13:34:01 crc kubenswrapper[4824]: E1124 13:34:01.533545 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n56fh5c5h554hd8h59hc6hffh675h57fh66fh659h646h59chbbh64bh659h56dh6bh5cch579h64fh55h566h58dh64h56dh545h5d4h8ch5fdh8fh5bdq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p6lkd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-558844779c-bw4xd_openstack(240e0890-6e9c-4b0b-95ce-786769479b06): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:34:01 crc kubenswrapper[4824]: E1124 13:34:01.536974 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-558844779c-bw4xd" podUID="240e0890-6e9c-4b0b-95ce-786769479b06" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.610602 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.784023 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4022c6cb-7074-46e7-a366-7005553fa97e-horizon-secret-key\") pod \"4022c6cb-7074-46e7-a366-7005553fa97e\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.784074 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4022c6cb-7074-46e7-a366-7005553fa97e-logs\") pod \"4022c6cb-7074-46e7-a366-7005553fa97e\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.784175 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd7mt\" (UniqueName: \"kubernetes.io/projected/4022c6cb-7074-46e7-a366-7005553fa97e-kube-api-access-pd7mt\") pod \"4022c6cb-7074-46e7-a366-7005553fa97e\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.784321 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-config-data\") pod \"4022c6cb-7074-46e7-a366-7005553fa97e\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.784363 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-scripts\") pod \"4022c6cb-7074-46e7-a366-7005553fa97e\" (UID: \"4022c6cb-7074-46e7-a366-7005553fa97e\") " Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.785202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4022c6cb-7074-46e7-a366-7005553fa97e-logs" (OuterVolumeSpecName: "logs") pod "4022c6cb-7074-46e7-a366-7005553fa97e" (UID: "4022c6cb-7074-46e7-a366-7005553fa97e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.785294 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-scripts" (OuterVolumeSpecName: "scripts") pod "4022c6cb-7074-46e7-a366-7005553fa97e" (UID: "4022c6cb-7074-46e7-a366-7005553fa97e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.785751 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-config-data" (OuterVolumeSpecName: "config-data") pod "4022c6cb-7074-46e7-a366-7005553fa97e" (UID: "4022c6cb-7074-46e7-a366-7005553fa97e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.795188 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4022c6cb-7074-46e7-a366-7005553fa97e-kube-api-access-pd7mt" (OuterVolumeSpecName: "kube-api-access-pd7mt") pod "4022c6cb-7074-46e7-a366-7005553fa97e" (UID: "4022c6cb-7074-46e7-a366-7005553fa97e"). InnerVolumeSpecName "kube-api-access-pd7mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.797039 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4022c6cb-7074-46e7-a366-7005553fa97e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4022c6cb-7074-46e7-a366-7005553fa97e" (UID: "4022c6cb-7074-46e7-a366-7005553fa97e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.885884 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.885913 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4022c6cb-7074-46e7-a366-7005553fa97e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.885922 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4022c6cb-7074-46e7-a366-7005553fa97e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.885934 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4022c6cb-7074-46e7-a366-7005553fa97e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:01 crc kubenswrapper[4824]: I1124 13:34:01.885943 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd7mt\" (UniqueName: \"kubernetes.io/projected/4022c6cb-7074-46e7-a366-7005553fa97e-kube-api-access-pd7mt\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:02 crc kubenswrapper[4824]: I1124 13:34:02.001991 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bccbd8f7f-rjdmz" Nov 24 13:34:02 crc kubenswrapper[4824]: I1124 13:34:02.001786 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bccbd8f7f-rjdmz" event={"ID":"4022c6cb-7074-46e7-a366-7005553fa97e","Type":"ContainerDied","Data":"9f6827e17ee5e398d973d7f21f79dbe6a1d252ba2ca7ac4d6060514fc512700c"} Nov 24 13:34:02 crc kubenswrapper[4824]: E1124 13:34:02.005606 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-8wqh7" podUID="aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" Nov 24 13:34:02 crc kubenswrapper[4824]: I1124 13:34:02.089927 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6bccbd8f7f-rjdmz"] Nov 24 13:34:02 crc kubenswrapper[4824]: I1124 13:34:02.105176 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6bccbd8f7f-rjdmz"] Nov 24 13:34:03 crc kubenswrapper[4824]: I1124 13:34:03.028698 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4022c6cb-7074-46e7-a366-7005553fa97e" path="/var/lib/kubelet/pods/4022c6cb-7074-46e7-a366-7005553fa97e/volumes" Nov 24 13:34:03 crc kubenswrapper[4824]: I1124 13:34:03.640782 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-74w28" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: i/o timeout" Nov 24 13:34:03 crc kubenswrapper[4824]: I1124 13:34:03.641440 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:34:03 crc kubenswrapper[4824]: E1124 13:34:03.955075 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 24 13:34:03 crc kubenswrapper[4824]: E1124 13:34:03.955392 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc5hcdh7fh548hf6hd5h595h66fhfbh5bh575h57h564h5cbh678h549h58bh586h9ch655hc4h658h9dh55bh7bh5b7h576h588h5f4h584h78h57cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zfxll,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-8857bfdc9-phjfh_openstack(317fbcea-64fb-4132-b028-4b760b98f8b3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:34:03 crc kubenswrapper[4824]: E1124 13:34:03.963213 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-8857bfdc9-phjfh" podUID="317fbcea-64fb-4132-b028-4b760b98f8b3" Nov 24 13:34:07 crc kubenswrapper[4824]: I1124 13:34:07.044467 4824 generic.go:334] "Generic (PLEG): container finished" podID="8e88a010-ca80-44fc-bd45-d741e3acdc47" containerID="4fe10b3cae19503781c7eea75a17365051a790c7b797f76de1ca4c3a7db834a3" exitCode=0 Nov 24 13:34:07 crc kubenswrapper[4824]: I1124 13:34:07.044617 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-95tmx" event={"ID":"8e88a010-ca80-44fc-bd45-d741e3acdc47","Type":"ContainerDied","Data":"4fe10b3cae19503781c7eea75a17365051a790c7b797f76de1ca4c3a7db834a3"} Nov 24 13:34:08 crc kubenswrapper[4824]: I1124 13:34:08.641412 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-74w28" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: i/o timeout" Nov 24 13:34:08 crc kubenswrapper[4824]: I1124 13:34:08.764547 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:34:08 crc kubenswrapper[4824]: I1124 13:34:08.764597 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:34:12 crc kubenswrapper[4824]: I1124 13:34:12.167285 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:12 crc kubenswrapper[4824]: I1124 13:34:12.167770 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.642286 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-74w28" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: i/o timeout" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.794032 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.799155 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.807999 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.810453 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.836739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-sb\") pod \"429c7611-6a69-43a2-8047-d46b48f48447\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.836784 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-combined-ca-bundle\") pod \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.836825 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.849319 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-config\") pod \"429c7611-6a69-43a2-8047-d46b48f48447\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.849708 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-dns-svc\") pod \"429c7611-6a69-43a2-8047-d46b48f48447\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.849777 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wstfq\" (UniqueName: \"kubernetes.io/projected/429c7611-6a69-43a2-8047-d46b48f48447-kube-api-access-wstfq\") pod \"429c7611-6a69-43a2-8047-d46b48f48447\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.849866 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69q58\" (UniqueName: \"kubernetes.io/projected/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-kube-api-access-69q58\") pod \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.849901 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-config-data\") pod \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.849920 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-combined-ca-bundle\") pod \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.849949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-httpd-run\") pod \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.849997 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.850021 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-config-data\") pod \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.851997 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-logs\") pod \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.852032 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-logs\") pod \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.852054 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-scripts\") pod \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.852120 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-httpd-run\") pod \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.852164 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwmj8\" (UniqueName: \"kubernetes.io/projected/3303f1e8-4944-430e-93e6-31bafe0e8c6b-kube-api-access-wwmj8\") pod \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\" (UID: \"3303f1e8-4944-430e-93e6-31bafe0e8c6b\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.852180 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-scripts\") pod \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\" (UID: \"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.857595 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-nb\") pod \"429c7611-6a69-43a2-8047-d46b48f48447\" (UID: \"429c7611-6a69-43a2-8047-d46b48f48447\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.859387 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-logs" (OuterVolumeSpecName: "logs") pod "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" (UID: "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.862688 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-logs" (OuterVolumeSpecName: "logs") pod "3303f1e8-4944-430e-93e6-31bafe0e8c6b" (UID: "3303f1e8-4944-430e-93e6-31bafe0e8c6b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.878902 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3303f1e8-4944-430e-93e6-31bafe0e8c6b" (UID: "3303f1e8-4944-430e-93e6-31bafe0e8c6b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.883748 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" (UID: "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.885341 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-kube-api-access-69q58" (OuterVolumeSpecName: "kube-api-access-69q58") pod "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" (UID: "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe"). InnerVolumeSpecName "kube-api-access-69q58". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.888451 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-scripts" (OuterVolumeSpecName: "scripts") pod "3303f1e8-4944-430e-93e6-31bafe0e8c6b" (UID: "3303f1e8-4944-430e-93e6-31bafe0e8c6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.890505 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "3303f1e8-4944-430e-93e6-31bafe0e8c6b" (UID: "3303f1e8-4944-430e-93e6-31bafe0e8c6b"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.890962 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" (UID: "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.891135 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-scripts" (OuterVolumeSpecName: "scripts") pod "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" (UID: "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.892407 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/429c7611-6a69-43a2-8047-d46b48f48447-kube-api-access-wstfq" (OuterVolumeSpecName: "kube-api-access-wstfq") pod "429c7611-6a69-43a2-8047-d46b48f48447" (UID: "429c7611-6a69-43a2-8047-d46b48f48447"). InnerVolumeSpecName "kube-api-access-wstfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.925579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" (UID: "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.931739 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3303f1e8-4944-430e-93e6-31bafe0e8c6b-kube-api-access-wwmj8" (OuterVolumeSpecName: "kube-api-access-wwmj8") pod "3303f1e8-4944-430e-93e6-31bafe0e8c6b" (UID: "3303f1e8-4944-430e-93e6-31bafe0e8c6b"). InnerVolumeSpecName "kube-api-access-wwmj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.949594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-config" (OuterVolumeSpecName: "config") pod "429c7611-6a69-43a2-8047-d46b48f48447" (UID: "429c7611-6a69-43a2-8047-d46b48f48447"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.949928 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "429c7611-6a69-43a2-8047-d46b48f48447" (UID: "429c7611-6a69-43a2-8047-d46b48f48447"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.949991 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3303f1e8-4944-430e-93e6-31bafe0e8c6b" (UID: "3303f1e8-4944-430e-93e6-31bafe0e8c6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.952270 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-config-data" (OuterVolumeSpecName: "config-data") pod "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" (UID: "ecfd309b-d902-42aa-b49b-2a0a5dacfdbe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.962310 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "429c7611-6a69-43a2-8047-d46b48f48447" (UID: "429c7611-6a69-43a2-8047-d46b48f48447"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.962477 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-scripts\") pod \"240e0890-6e9c-4b0b-95ce-786769479b06\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.962530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-config-data\") pod \"240e0890-6e9c-4b0b-95ce-786769479b06\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.962567 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/240e0890-6e9c-4b0b-95ce-786769479b06-logs\") pod \"240e0890-6e9c-4b0b-95ce-786769479b06\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.962671 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6lkd\" (UniqueName: \"kubernetes.io/projected/240e0890-6e9c-4b0b-95ce-786769479b06-kube-api-access-p6lkd\") pod \"240e0890-6e9c-4b0b-95ce-786769479b06\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.962726 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/240e0890-6e9c-4b0b-95ce-786769479b06-horizon-secret-key\") pod \"240e0890-6e9c-4b0b-95ce-786769479b06\" (UID: \"240e0890-6e9c-4b0b-95ce-786769479b06\") " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.963589 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/240e0890-6e9c-4b0b-95ce-786769479b06-logs" (OuterVolumeSpecName: "logs") pod "240e0890-6e9c-4b0b-95ce-786769479b06" (UID: "240e0890-6e9c-4b0b-95ce-786769479b06"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.963684 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-scripts" (OuterVolumeSpecName: "scripts") pod "240e0890-6e9c-4b0b-95ce-786769479b06" (UID: "240e0890-6e9c-4b0b-95ce-786769479b06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.963704 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-config-data" (OuterVolumeSpecName: "config-data") pod "240e0890-6e9c-4b0b-95ce-786769479b06" (UID: "240e0890-6e9c-4b0b-95ce-786769479b06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.963969 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.963987 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.963997 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964006 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/240e0890-6e9c-4b0b-95ce-786769479b06-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964016 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/240e0890-6e9c-4b0b-95ce-786769479b06-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964027 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wstfq\" (UniqueName: \"kubernetes.io/projected/429c7611-6a69-43a2-8047-d46b48f48447-kube-api-access-wstfq\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964038 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69q58\" (UniqueName: \"kubernetes.io/projected/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-kube-api-access-69q58\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964046 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964055 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964064 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964085 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964094 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964102 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964110 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964118 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3303f1e8-4944-430e-93e6-31bafe0e8c6b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964126 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwmj8\" (UniqueName: \"kubernetes.io/projected/3303f1e8-4944-430e-93e6-31bafe0e8c6b-kube-api-access-wwmj8\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964135 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964143 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964151 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.964160 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.966300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/240e0890-6e9c-4b0b-95ce-786769479b06-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "240e0890-6e9c-4b0b-95ce-786769479b06" (UID: "240e0890-6e9c-4b0b-95ce-786769479b06"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.972317 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "429c7611-6a69-43a2-8047-d46b48f48447" (UID: "429c7611-6a69-43a2-8047-d46b48f48447"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.978153 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/240e0890-6e9c-4b0b-95ce-786769479b06-kube-api-access-p6lkd" (OuterVolumeSpecName: "kube-api-access-p6lkd") pod "240e0890-6e9c-4b0b-95ce-786769479b06" (UID: "240e0890-6e9c-4b0b-95ce-786769479b06"). InnerVolumeSpecName "kube-api-access-p6lkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.983575 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-config-data" (OuterVolumeSpecName: "config-data") pod "3303f1e8-4944-430e-93e6-31bafe0e8c6b" (UID: "3303f1e8-4944-430e-93e6-31bafe0e8c6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.992295 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 24 13:34:13 crc kubenswrapper[4824]: I1124 13:34:13.994078 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.065938 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.065967 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3303f1e8-4944-430e-93e6-31bafe0e8c6b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.065979 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6lkd\" (UniqueName: \"kubernetes.io/projected/240e0890-6e9c-4b0b-95ce-786769479b06-kube-api-access-p6lkd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.065991 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/240e0890-6e9c-4b0b-95ce-786769479b06-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.066100 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.066510 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/429c7611-6a69-43a2-8047-d46b48f48447-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.126041 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-558844779c-bw4xd" event={"ID":"240e0890-6e9c-4b0b-95ce-786769479b06","Type":"ContainerDied","Data":"b4e357bf17f6920b3b15a07765e563542f08c534c433ab0e7f51da9475a79e5e"} Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.126208 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-558844779c-bw4xd" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.136063 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-74w28" event={"ID":"429c7611-6a69-43a2-8047-d46b48f48447","Type":"ContainerDied","Data":"81f14dd73e3f1b0e26b8ffb4b8c9526d8f8c77b94beee9b76eda0d9f7d12b90f"} Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.136095 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-74w28" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.136113 4824 scope.go:117] "RemoveContainer" containerID="0eacd6c6d75a66c6c694c6209d042d8d8cefc528e587d8758df40d15b776d18d" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.138688 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3303f1e8-4944-430e-93e6-31bafe0e8c6b","Type":"ContainerDied","Data":"9b7537adc8776d2c889ac3ff94af26c9c0ac809ea5b1170598f6a9092caa8855"} Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.138735 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.147775 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ecfd309b-d902-42aa-b49b-2a0a5dacfdbe","Type":"ContainerDied","Data":"26f78abc9b893dd7339573f4b49fbc7ba563f6d29d63b4ea9eb251e2e0c2032c"} Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.147957 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.197057 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-558844779c-bw4xd"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.198500 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-558844779c-bw4xd"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.210731 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-74w28"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.220968 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-74w28"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.223826 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.230953 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.237467 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.244939 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.253760 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:14 crc kubenswrapper[4824]: E1124 13:34:14.254208 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerName="glance-httpd" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254222 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerName="glance-httpd" Nov 24 13:34:14 crc kubenswrapper[4824]: E1124 13:34:14.254235 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerName="glance-httpd" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254241 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerName="glance-httpd" Nov 24 13:34:14 crc kubenswrapper[4824]: E1124 13:34:14.254252 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerName="glance-log" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254258 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerName="glance-log" Nov 24 13:34:14 crc kubenswrapper[4824]: E1124 13:34:14.254282 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="init" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254288 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="init" Nov 24 13:34:14 crc kubenswrapper[4824]: E1124 13:34:14.254311 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254317 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" Nov 24 13:34:14 crc kubenswrapper[4824]: E1124 13:34:14.254328 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerName="glance-log" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254334 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerName="glance-log" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254489 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254500 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerName="glance-log" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254510 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" containerName="glance-httpd" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254527 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerName="glance-log" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.254539 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" containerName="glance-httpd" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.255403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.257221 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-p7tpr" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.257487 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.257678 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.257969 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.265521 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.274224 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.276937 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.279398 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.281838 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.307584 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.373777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-scripts\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374140 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374401 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374446 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374483 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4qmp\" (UniqueName: \"kubernetes.io/projected/36a320b3-eeb2-4400-8d0e-27360ab167ee-kube-api-access-q4qmp\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-config-data\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374523 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-logs\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374565 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374597 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-logs\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fh4q\" (UniqueName: \"kubernetes.io/projected/6b60d128-15e0-4646-ad33-c562d28aedbc-kube-api-access-4fh4q\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.374692 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: E1124 13:34:14.415078 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 24 13:34:14 crc kubenswrapper[4824]: E1124 13:34:14.415210 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf6h56dh5cch548hf6h5d5h68fh55fh694h5d5h679h5ch55fh557h5b7h8ch54ch5ffh699h5b8hffhd8hd5h7bh55fh65bh5bbh79h65bh56bh5cdh694q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-flfrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(597521ee-8310-4c63-bdd2-aeef878fb8b9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.435148 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.444063 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-95tmx" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.476722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-combined-ca-bundle\") pod \"8e88a010-ca80-44fc-bd45-d741e3acdc47\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.476799 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/317fbcea-64fb-4132-b028-4b760b98f8b3-horizon-secret-key\") pod \"317fbcea-64fb-4132-b028-4b760b98f8b3\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.476864 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-config\") pod \"8e88a010-ca80-44fc-bd45-d741e3acdc47\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.476973 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l59l\" (UniqueName: \"kubernetes.io/projected/8e88a010-ca80-44fc-bd45-d741e3acdc47-kube-api-access-5l59l\") pod \"8e88a010-ca80-44fc-bd45-d741e3acdc47\" (UID: \"8e88a010-ca80-44fc-bd45-d741e3acdc47\") " Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.477027 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-scripts\") pod \"317fbcea-64fb-4132-b028-4b760b98f8b3\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.477041 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-config-data\") pod \"317fbcea-64fb-4132-b028-4b760b98f8b3\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.477089 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/317fbcea-64fb-4132-b028-4b760b98f8b3-logs\") pod \"317fbcea-64fb-4132-b028-4b760b98f8b3\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.477508 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfxll\" (UniqueName: \"kubernetes.io/projected/317fbcea-64fb-4132-b028-4b760b98f8b3-kube-api-access-zfxll\") pod \"317fbcea-64fb-4132-b028-4b760b98f8b3\" (UID: \"317fbcea-64fb-4132-b028-4b760b98f8b3\") " Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.477693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-logs\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.477717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fh4q\" (UniqueName: \"kubernetes.io/projected/6b60d128-15e0-4646-ad33-c562d28aedbc-kube-api-access-4fh4q\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.477763 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.478132 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-scripts\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.478635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.478654 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.478878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.478942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.478962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479074 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-scripts" (OuterVolumeSpecName: "scripts") pod "317fbcea-64fb-4132-b028-4b760b98f8b3" (UID: "317fbcea-64fb-4132-b028-4b760b98f8b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479125 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479324 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4qmp\" (UniqueName: \"kubernetes.io/projected/36a320b3-eeb2-4400-8d0e-27360ab167ee-kube-api-access-q4qmp\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479347 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-config-data\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-logs\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/317fbcea-64fb-4132-b028-4b760b98f8b3-logs" (OuterVolumeSpecName: "logs") pod "317fbcea-64fb-4132-b028-4b760b98f8b3" (UID: "317fbcea-64fb-4132-b028-4b760b98f8b3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479438 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.479994 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-config-data" (OuterVolumeSpecName: "config-data") pod "317fbcea-64fb-4132-b028-4b760b98f8b3" (UID: "317fbcea-64fb-4132-b028-4b760b98f8b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.487102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-config-data\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.487510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-logs\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.487762 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.510459 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.513478 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-logs\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.516217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.516659 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-scripts\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.519327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.519594 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.527503 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.527792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.536621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.544521 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e88a010-ca80-44fc-bd45-d741e3acdc47-kube-api-access-5l59l" (OuterVolumeSpecName: "kube-api-access-5l59l") pod "8e88a010-ca80-44fc-bd45-d741e3acdc47" (UID: "8e88a010-ca80-44fc-bd45-d741e3acdc47"). InnerVolumeSpecName "kube-api-access-5l59l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.545714 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.546513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.547905 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.551651 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fh4q\" (UniqueName: \"kubernetes.io/projected/6b60d128-15e0-4646-ad33-c562d28aedbc-kube-api-access-4fh4q\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.555432 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/317fbcea-64fb-4132-b028-4b760b98f8b3-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "317fbcea-64fb-4132-b028-4b760b98f8b3" (UID: "317fbcea-64fb-4132-b028-4b760b98f8b3"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.555764 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/317fbcea-64fb-4132-b028-4b760b98f8b3-kube-api-access-zfxll" (OuterVolumeSpecName: "kube-api-access-zfxll") pod "317fbcea-64fb-4132-b028-4b760b98f8b3" (UID: "317fbcea-64fb-4132-b028-4b760b98f8b3"). InnerVolumeSpecName "kube-api-access-zfxll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.557855 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " pod="openstack/glance-default-external-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.563945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e88a010-ca80-44fc-bd45-d741e3acdc47" (UID: "8e88a010-ca80-44fc-bd45-d741e3acdc47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.565544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4qmp\" (UniqueName: \"kubernetes.io/projected/36a320b3-eeb2-4400-8d0e-27360ab167ee-kube-api-access-q4qmp\") pod \"glance-default-internal-api-0\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.568702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-config" (OuterVolumeSpecName: "config") pod "8e88a010-ca80-44fc-bd45-d741e3acdc47" (UID: "8e88a010-ca80-44fc-bd45-d741e3acdc47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.571294 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.581084 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/317fbcea-64fb-4132-b028-4b760b98f8b3-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.581104 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.581115 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l59l\" (UniqueName: \"kubernetes.io/projected/8e88a010-ca80-44fc-bd45-d741e3acdc47-kube-api-access-5l59l\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.581124 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/317fbcea-64fb-4132-b028-4b760b98f8b3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.581161 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/317fbcea-64fb-4132-b028-4b760b98f8b3-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.581171 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfxll\" (UniqueName: \"kubernetes.io/projected/317fbcea-64fb-4132-b028-4b760b98f8b3-kube-api-access-zfxll\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.581179 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e88a010-ca80-44fc-bd45-d741e3acdc47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:14 crc kubenswrapper[4824]: I1124 13:34:14.627076 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.019859 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="240e0890-6e9c-4b0b-95ce-786769479b06" path="/var/lib/kubelet/pods/240e0890-6e9c-4b0b-95ce-786769479b06/volumes" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.020441 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3303f1e8-4944-430e-93e6-31bafe0e8c6b" path="/var/lib/kubelet/pods/3303f1e8-4944-430e-93e6-31bafe0e8c6b/volumes" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.021385 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="429c7611-6a69-43a2-8047-d46b48f48447" path="/var/lib/kubelet/pods/429c7611-6a69-43a2-8047-d46b48f48447/volumes" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.024643 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecfd309b-d902-42aa-b49b-2a0a5dacfdbe" path="/var/lib/kubelet/pods/ecfd309b-d902-42aa-b49b-2a0a5dacfdbe/volumes" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.162260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-95tmx" event={"ID":"8e88a010-ca80-44fc-bd45-d741e3acdc47","Type":"ContainerDied","Data":"05ac447074bc5fa703ab08b1b9118885c9bc27fbd76f38087b7a05ae5bb3430f"} Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.162296 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05ac447074bc5fa703ab08b1b9118885c9bc27fbd76f38087b7a05ae5bb3430f" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.162354 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-95tmx" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.163995 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8857bfdc9-phjfh" event={"ID":"317fbcea-64fb-4132-b028-4b760b98f8b3","Type":"ContainerDied","Data":"b82e79ad288629d9b6cd95ca2b66618a6ba21871b6feb7b9cb424f85f9911bd0"} Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.164046 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8857bfdc9-phjfh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.211383 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8857bfdc9-phjfh"] Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.221795 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8857bfdc9-phjfh"] Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.749913 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r9wlh"] Nov 24 13:34:15 crc kubenswrapper[4824]: E1124 13:34:15.750485 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e88a010-ca80-44fc-bd45-d741e3acdc47" containerName="neutron-db-sync" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.750752 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e88a010-ca80-44fc-bd45-d741e3acdc47" containerName="neutron-db-sync" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.750998 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e88a010-ca80-44fc-bd45-d741e3acdc47" containerName="neutron-db-sync" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.753107 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.776854 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r9wlh"] Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.805874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-svc\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.805973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6rzg\" (UniqueName: \"kubernetes.io/projected/984c9282-012e-443d-b80a-89133d02dcde-kube-api-access-g6rzg\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.806008 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-config\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.806043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.806078 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.806106 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.907969 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6rzg\" (UniqueName: \"kubernetes.io/projected/984c9282-012e-443d-b80a-89133d02dcde-kube-api-access-g6rzg\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.908035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-config\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.908089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.908136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.908170 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.908248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-svc\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.909287 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-svc\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.909387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-config\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.909420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.909899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.909980 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:15 crc kubenswrapper[4824]: I1124 13:34:15.934933 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6rzg\" (UniqueName: \"kubernetes.io/projected/984c9282-012e-443d-b80a-89133d02dcde-kube-api-access-g6rzg\") pod \"dnsmasq-dns-55f844cf75-r9wlh\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.026672 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7d494558c4-pqqd8"] Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.028291 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.031359 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.031427 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.031429 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.031395 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fmdf7" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.042448 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d494558c4-pqqd8"] Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.083455 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.111128 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-httpd-config\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.111177 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-combined-ca-bundle\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.111207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-config\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.111223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-ovndb-tls-certs\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.111251 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g4gw\" (UniqueName: \"kubernetes.io/projected/a5c8cb45-5223-4339-a957-80ec09aad094-kube-api-access-7g4gw\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.212994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-combined-ca-bundle\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.213051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-config\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.213071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-ovndb-tls-certs\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.213095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g4gw\" (UniqueName: \"kubernetes.io/projected/a5c8cb45-5223-4339-a957-80ec09aad094-kube-api-access-7g4gw\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.213193 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-httpd-config\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.217088 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-config\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.218004 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-httpd-config\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.221280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-combined-ca-bundle\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.228400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-ovndb-tls-certs\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.229343 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g4gw\" (UniqueName: \"kubernetes.io/projected/a5c8cb45-5223-4339-a957-80ec09aad094-kube-api-access-7g4gw\") pod \"neutron-7d494558c4-pqqd8\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.349175 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.472755 4824 scope.go:117] "RemoveContainer" containerID="27a099e0535ad549e5b39e4258eca640dc27d8c79443cd5ceccb564ff76634b1" Nov 24 13:34:16 crc kubenswrapper[4824]: E1124 13:34:16.512308 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 24 13:34:16 crc kubenswrapper[4824]: E1124 13:34:16.512454 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5dpm7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-69zcz_openstack(b1af3874-3566-4785-a5fe-13375aa658c3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 13:34:16 crc kubenswrapper[4824]: E1124 13:34:16.514207 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-69zcz" podUID="b1af3874-3566-4785-a5fe-13375aa658c3" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.654705 4824 scope.go:117] "RemoveContainer" containerID="a94b804e071c1782ef5b773b77d9f36a285e09bd557397e071100fbe6777b31d" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.831276 4824 scope.go:117] "RemoveContainer" containerID="0867909c9e672d429b921035060badefdb63e7e95df94652cbbe24560f991555" Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.932552 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-959bb6dd8-8pcrj"] Nov 24 13:34:16 crc kubenswrapper[4824]: I1124 13:34:16.932729 4824 scope.go:117] "RemoveContainer" containerID="fc8974a7cf2e902af051b1559277aecae1a57a482335156560c839bcd1a2d5f6" Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.039177 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="317fbcea-64fb-4132-b028-4b760b98f8b3" path="/var/lib/kubelet/pods/317fbcea-64fb-4132-b028-4b760b98f8b3/volumes" Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.083371 4824 scope.go:117] "RemoveContainer" containerID="b629cf93ec1fb2d48832ab8317c67ddf23fe380fbec6a5784ef5de4711a8d167" Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.091215 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c96f4646d-wsbgv"] Nov 24 13:34:17 crc kubenswrapper[4824]: W1124 13:34:17.123240 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc2d2fae_b656_464e_b1a8_89b4c5160792.slice/crio-95426b1d4e810601f0275b28c86b9c913b291bd7d431bf89c1f4441239ce5cbd WatchSource:0}: Error finding container 95426b1d4e810601f0275b28c86b9c913b291bd7d431bf89c1f4441239ce5cbd: Status 404 returned error can't find the container with id 95426b1d4e810601f0275b28c86b9c913b291bd7d431bf89c1f4441239ce5cbd Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.183196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-trpzx" event={"ID":"0781aef7-bf19-478e-931f-a10a3dcd586d","Type":"ContainerStarted","Data":"9362b01e70cca8836c8dea1c873faa494ce0ceebc8d74d961b4000e0192cc1b2"} Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.191084 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c96f4646d-wsbgv" event={"ID":"bc2d2fae-b656-464e-b1a8-89b4c5160792","Type":"ContainerStarted","Data":"95426b1d4e810601f0275b28c86b9c913b291bd7d431bf89c1f4441239ce5cbd"} Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.192992 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-959bb6dd8-8pcrj" event={"ID":"5828a51f-d146-44ea-b8ea-423e017939d8","Type":"ContainerStarted","Data":"2e5ac11b54ef6a28e4b3000958a43ce7a367ba462511dbd5ae27609c6fd6fde6"} Nov 24 13:34:17 crc kubenswrapper[4824]: E1124 13:34:17.205947 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-69zcz" podUID="b1af3874-3566-4785-a5fe-13375aa658c3" Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.207998 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-trpzx" podStartSLOduration=3.703443404 podStartE2EDuration="42.207978185s" podCreationTimestamp="2025-11-24 13:33:35 +0000 UTC" firstStartedPulling="2025-11-24 13:33:37.879170088 +0000 UTC m=+1159.518709398" lastFinishedPulling="2025-11-24 13:34:16.383704869 +0000 UTC m=+1198.023244179" observedRunningTime="2025-11-24 13:34:17.19924169 +0000 UTC m=+1198.838781000" watchObservedRunningTime="2025-11-24 13:34:17.207978185 +0000 UTC m=+1198.847517495" Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.290349 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nhzcv"] Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.444000 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r9wlh"] Nov 24 13:34:17 crc kubenswrapper[4824]: W1124 13:34:17.467120 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod984c9282_012e_443d_b80a_89133d02dcde.slice/crio-e896deeeac3e748d459f4b76fa9db8a670ee3ecc3906a06855128fd14dcc9865 WatchSource:0}: Error finding container e896deeeac3e748d459f4b76fa9db8a670ee3ecc3906a06855128fd14dcc9865: Status 404 returned error can't find the container with id e896deeeac3e748d459f4b76fa9db8a670ee3ecc3906a06855128fd14dcc9865 Nov 24 13:34:17 crc kubenswrapper[4824]: I1124 13:34:17.732089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:34:17 crc kubenswrapper[4824]: W1124 13:34:17.765931 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36a320b3_eeb2_4400_8d0e_27360ab167ee.slice/crio-029e8aada7d7da37038c43aa7bea540f12a8ba8032df61ee7851bc6f481d1f25 WatchSource:0}: Error finding container 029e8aada7d7da37038c43aa7bea540f12a8ba8032df61ee7851bc6f481d1f25: Status 404 returned error can't find the container with id 029e8aada7d7da37038c43aa7bea540f12a8ba8032df61ee7851bc6f481d1f25 Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.193419 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d494558c4-pqqd8"] Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.226787 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-959bb6dd8-8pcrj" event={"ID":"5828a51f-d146-44ea-b8ea-423e017939d8","Type":"ContainerStarted","Data":"da6f9a728a58a013b201b3a4e4041e55f6270f764a17d3d9b229efa34f13bb7b"} Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.232476 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nhzcv" event={"ID":"4735a096-fa3b-4a42-8ff1-fb1f92b2e953","Type":"ContainerStarted","Data":"96d6e74afc275b4335f284c28a31f9d22e683851773e95ed6c8b4c1090eda72d"} Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.232547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nhzcv" event={"ID":"4735a096-fa3b-4a42-8ff1-fb1f92b2e953","Type":"ContainerStarted","Data":"3fe2986614d7e2a51bfe3d7bb6bb93830334243ef31af9c9ff86492280677e4b"} Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.235045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"36a320b3-eeb2-4400-8d0e-27360ab167ee","Type":"ContainerStarted","Data":"029e8aada7d7da37038c43aa7bea540f12a8ba8032df61ee7851bc6f481d1f25"} Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.238189 4824 generic.go:334] "Generic (PLEG): container finished" podID="984c9282-012e-443d-b80a-89133d02dcde" containerID="5932ef8564ca81c459a1d6855d567023d3cce56ed33f9014d0d06deb77e7fb43" exitCode=0 Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.238213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" event={"ID":"984c9282-012e-443d-b80a-89133d02dcde","Type":"ContainerDied","Data":"5932ef8564ca81c459a1d6855d567023d3cce56ed33f9014d0d06deb77e7fb43"} Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.238240 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" event={"ID":"984c9282-012e-443d-b80a-89133d02dcde","Type":"ContainerStarted","Data":"e896deeeac3e748d459f4b76fa9db8a670ee3ecc3906a06855128fd14dcc9865"} Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.287082 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nhzcv" podStartSLOduration=24.287047821 podStartE2EDuration="24.287047821s" podCreationTimestamp="2025-11-24 13:33:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:18.257930291 +0000 UTC m=+1199.897469601" watchObservedRunningTime="2025-11-24 13:34:18.287047821 +0000 UTC m=+1199.926587131" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.493317 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-798f5d656f-r9xcp"] Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.497755 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.500160 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.500350 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.515702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-798f5d656f-r9xcp"] Nov 24 13:34:18 crc kubenswrapper[4824]: W1124 13:34:18.536487 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5c8cb45_5223_4339_a957_80ec09aad094.slice/crio-bf28b5a6d78804dec27ada4b698f71a36861860153318a7604d2ff8fa3f02def WatchSource:0}: Error finding container bf28b5a6d78804dec27ada4b698f71a36861860153318a7604d2ff8fa3f02def: Status 404 returned error can't find the container with id bf28b5a6d78804dec27ada4b698f71a36861860153318a7604d2ff8fa3f02def Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.622016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-public-tls-certs\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.622388 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-internal-tls-certs\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.622532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-combined-ca-bundle\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.622888 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-config\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.623042 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-ovndb-tls-certs\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.623158 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfbwz\" (UniqueName: \"kubernetes.io/projected/bf77fc79-5c9e-4688-be73-b390bef539fd-kube-api-access-tfbwz\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.623252 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-httpd-config\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.643308 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-74w28" podUID="429c7611-6a69-43a2-8047-d46b48f48447" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: i/o timeout" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.653029 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.725209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-ovndb-tls-certs\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.725251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfbwz\" (UniqueName: \"kubernetes.io/projected/bf77fc79-5c9e-4688-be73-b390bef539fd-kube-api-access-tfbwz\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.725286 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-httpd-config\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.725326 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-public-tls-certs\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.725391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-internal-tls-certs\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.725414 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-combined-ca-bundle\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.725456 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-config\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.730382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-internal-tls-certs\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.732077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-combined-ca-bundle\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.732161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-config\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.732621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-httpd-config\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.736086 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-public-tls-certs\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.741603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf77fc79-5c9e-4688-be73-b390bef539fd-ovndb-tls-certs\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.743108 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfbwz\" (UniqueName: \"kubernetes.io/projected/bf77fc79-5c9e-4688-be73-b390bef539fd-kube-api-access-tfbwz\") pod \"neutron-798f5d656f-r9xcp\" (UID: \"bf77fc79-5c9e-4688-be73-b390bef539fd\") " pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:18 crc kubenswrapper[4824]: I1124 13:34:18.791711 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.323533 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" event={"ID":"984c9282-012e-443d-b80a-89133d02dcde","Type":"ContainerStarted","Data":"58e8551b4a63b882692d2d8d91f967797db29235af2bf39d0874dfd9fd229779"} Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.325270 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.346550 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8wqh7" event={"ID":"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce","Type":"ContainerStarted","Data":"6761b3024842943636b1639e6b9bc2ce39c1fa754d686e618def474225da2dd0"} Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.364032 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" podStartSLOduration=4.364016503 podStartE2EDuration="4.364016503s" podCreationTimestamp="2025-11-24 13:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:19.349489339 +0000 UTC m=+1200.989028649" watchObservedRunningTime="2025-11-24 13:34:19.364016503 +0000 UTC m=+1201.003555813" Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.383029 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-8wqh7" podStartSLOduration=3.817094749 podStartE2EDuration="44.383012772s" podCreationTimestamp="2025-11-24 13:33:35 +0000 UTC" firstStartedPulling="2025-11-24 13:33:37.119067604 +0000 UTC m=+1158.758606914" lastFinishedPulling="2025-11-24 13:34:17.684985627 +0000 UTC m=+1199.324524937" observedRunningTime="2025-11-24 13:34:19.382262733 +0000 UTC m=+1201.021802033" watchObservedRunningTime="2025-11-24 13:34:19.383012772 +0000 UTC m=+1201.022552082" Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.386995 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-959bb6dd8-8pcrj" event={"ID":"5828a51f-d146-44ea-b8ea-423e017939d8","Type":"ContainerStarted","Data":"f0eb7c3f4f829b0e7b8112bc730b12912e7eb168f9eb8c9618b6934afbce96bd"} Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.395014 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b60d128-15e0-4646-ad33-c562d28aedbc","Type":"ContainerStarted","Data":"b40991aa3544468f25ea0be0078985ffc0980bc424f1240478a3d08510a2bf6a"} Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.406121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d494558c4-pqqd8" event={"ID":"a5c8cb45-5223-4339-a957-80ec09aad094","Type":"ContainerStarted","Data":"c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea"} Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.406167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d494558c4-pqqd8" event={"ID":"a5c8cb45-5223-4339-a957-80ec09aad094","Type":"ContainerStarted","Data":"bf28b5a6d78804dec27ada4b698f71a36861860153318a7604d2ff8fa3f02def"} Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.415793 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-959bb6dd8-8pcrj" podStartSLOduration=29.882921566 podStartE2EDuration="30.415777696s" podCreationTimestamp="2025-11-24 13:33:49 +0000 UTC" firstStartedPulling="2025-11-24 13:34:16.951641514 +0000 UTC m=+1198.591180824" lastFinishedPulling="2025-11-24 13:34:17.484497644 +0000 UTC m=+1199.124036954" observedRunningTime="2025-11-24 13:34:19.411243269 +0000 UTC m=+1201.050782589" watchObservedRunningTime="2025-11-24 13:34:19.415777696 +0000 UTC m=+1201.055317006" Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.418046 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"597521ee-8310-4c63-bdd2-aeef878fb8b9","Type":"ContainerStarted","Data":"a6898c246a7e470e32264232cbde12bde5dfa4a4dd33ccaf1bb34d408b7cbdb6"} Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.425768 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"36a320b3-eeb2-4400-8d0e-27360ab167ee","Type":"ContainerStarted","Data":"3abafe77dceb2eda0505bc39ac04ec7eda6553623108aa3cdedfb8ec7c03da03"} Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.433663 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c96f4646d-wsbgv" event={"ID":"bc2d2fae-b656-464e-b1a8-89b4c5160792","Type":"ContainerStarted","Data":"8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933"} Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.577054 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-798f5d656f-r9xcp"] Nov 24 13:34:19 crc kubenswrapper[4824]: W1124 13:34:19.581368 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf77fc79_5c9e_4688_be73_b390bef539fd.slice/crio-e6774828c6293c05edec60d4a716699340861e4efe542f62cfb45ad5f3c8e826 WatchSource:0}: Error finding container e6774828c6293c05edec60d4a716699340861e4efe542f62cfb45ad5f3c8e826: Status 404 returned error can't find the container with id e6774828c6293c05edec60d4a716699340861e4efe542f62cfb45ad5f3c8e826 Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.673665 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:34:19 crc kubenswrapper[4824]: I1124 13:34:19.673729 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.470122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"36a320b3-eeb2-4400-8d0e-27360ab167ee","Type":"ContainerStarted","Data":"93a38e62aa44c86cd016c05ce04fc6b0cf9264567f9ff24a5340836da177bed2"} Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.483425 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c96f4646d-wsbgv" event={"ID":"bc2d2fae-b656-464e-b1a8-89b4c5160792","Type":"ContainerStarted","Data":"6ce765b74f266b883a095df6fbfd4992669f9774df16a85346b8570f4fc1ede0"} Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.486931 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b60d128-15e0-4646-ad33-c562d28aedbc","Type":"ContainerStarted","Data":"460c47f4638325b913b7972be871cd2beee6d07988efb6ce1b04ef745b1130d0"} Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.490477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-798f5d656f-r9xcp" event={"ID":"bf77fc79-5c9e-4688-be73-b390bef539fd","Type":"ContainerStarted","Data":"5e35ac3bc89bd97c2bca8b4cbe227a85985057aaed8ee2a4935a79336af813a4"} Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.490531 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-798f5d656f-r9xcp" event={"ID":"bf77fc79-5c9e-4688-be73-b390bef539fd","Type":"ContainerStarted","Data":"a495c6ff795d7125e070092799a1461f979d7d1aca22616c5fe6f86641cb22cf"} Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.490541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-798f5d656f-r9xcp" event={"ID":"bf77fc79-5c9e-4688-be73-b390bef539fd","Type":"ContainerStarted","Data":"e6774828c6293c05edec60d4a716699340861e4efe542f62cfb45ad5f3c8e826"} Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.490752 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.502792 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d494558c4-pqqd8" event={"ID":"a5c8cb45-5223-4339-a957-80ec09aad094","Type":"ContainerStarted","Data":"f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56"} Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.503339 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.515298 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.515273489 podStartE2EDuration="6.515273489s" podCreationTimestamp="2025-11-24 13:34:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:20.502072599 +0000 UTC m=+1202.141611909" watchObservedRunningTime="2025-11-24 13:34:20.515273489 +0000 UTC m=+1202.154812799" Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.532168 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-798f5d656f-r9xcp" podStartSLOduration=2.532153854 podStartE2EDuration="2.532153854s" podCreationTimestamp="2025-11-24 13:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:20.528949331 +0000 UTC m=+1202.168488641" watchObservedRunningTime="2025-11-24 13:34:20.532153854 +0000 UTC m=+1202.171693164" Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.597663 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c96f4646d-wsbgv" podStartSLOduration=31.03993379 podStartE2EDuration="31.5976428s" podCreationTimestamp="2025-11-24 13:33:49 +0000 UTC" firstStartedPulling="2025-11-24 13:34:17.127211945 +0000 UTC m=+1198.766751255" lastFinishedPulling="2025-11-24 13:34:17.684920965 +0000 UTC m=+1199.324460265" observedRunningTime="2025-11-24 13:34:20.593224836 +0000 UTC m=+1202.232764146" watchObservedRunningTime="2025-11-24 13:34:20.5976428 +0000 UTC m=+1202.237182110" Nov 24 13:34:20 crc kubenswrapper[4824]: I1124 13:34:20.671566 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7d494558c4-pqqd8" podStartSLOduration=4.6715484830000005 podStartE2EDuration="4.671548483s" podCreationTimestamp="2025-11-24 13:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:20.619194375 +0000 UTC m=+1202.258733685" watchObservedRunningTime="2025-11-24 13:34:20.671548483 +0000 UTC m=+1202.311087793" Nov 24 13:34:21 crc kubenswrapper[4824]: I1124 13:34:21.516343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b60d128-15e0-4646-ad33-c562d28aedbc","Type":"ContainerStarted","Data":"d1a92abb4c37ed9efbc60fa33188f3ff303d3d8678908e453a7329176942773b"} Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.539018 4824 generic.go:334] "Generic (PLEG): container finished" podID="0781aef7-bf19-478e-931f-a10a3dcd586d" containerID="9362b01e70cca8836c8dea1c873faa494ce0ceebc8d74d961b4000e0192cc1b2" exitCode=0 Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.539085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-trpzx" event={"ID":"0781aef7-bf19-478e-931f-a10a3dcd586d","Type":"ContainerDied","Data":"9362b01e70cca8836c8dea1c873faa494ce0ceebc8d74d961b4000e0192cc1b2"} Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.541631 4824 generic.go:334] "Generic (PLEG): container finished" podID="aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" containerID="6761b3024842943636b1639e6b9bc2ce39c1fa754d686e618def474225da2dd0" exitCode=0 Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.541656 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8wqh7" event={"ID":"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce","Type":"ContainerDied","Data":"6761b3024842943636b1639e6b9bc2ce39c1fa754d686e618def474225da2dd0"} Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.556885 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.556871282 podStartE2EDuration="10.556871282s" podCreationTimestamp="2025-11-24 13:34:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:21.542340266 +0000 UTC m=+1203.181879596" watchObservedRunningTime="2025-11-24 13:34:24.556871282 +0000 UTC m=+1206.196410592" Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.573795 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.573865 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.612870 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.626916 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.627215 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.627243 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.680142 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 13:34:24 crc kubenswrapper[4824]: I1124 13:34:24.684689 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 13:34:25 crc kubenswrapper[4824]: I1124 13:34:25.556339 4824 generic.go:334] "Generic (PLEG): container finished" podID="4735a096-fa3b-4a42-8ff1-fb1f92b2e953" containerID="96d6e74afc275b4335f284c28a31f9d22e683851773e95ed6c8b4c1090eda72d" exitCode=0 Nov 24 13:34:25 crc kubenswrapper[4824]: I1124 13:34:25.556384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nhzcv" event={"ID":"4735a096-fa3b-4a42-8ff1-fb1f92b2e953","Type":"ContainerDied","Data":"96d6e74afc275b4335f284c28a31f9d22e683851773e95ed6c8b4c1090eda72d"} Nov 24 13:34:25 crc kubenswrapper[4824]: I1124 13:34:25.557259 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:25 crc kubenswrapper[4824]: I1124 13:34:25.557405 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:34:25 crc kubenswrapper[4824]: I1124 13:34:25.557434 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:34:25 crc kubenswrapper[4824]: I1124 13:34:25.557445 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:26 crc kubenswrapper[4824]: I1124 13:34:26.084964 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:26 crc kubenswrapper[4824]: I1124 13:34:26.163912 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-ktmj4"] Nov 24 13:34:26 crc kubenswrapper[4824]: I1124 13:34:26.164410 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" podUID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" containerName="dnsmasq-dns" containerID="cri-o://8a39ea221e0a460c01cbe5046e500c3a42d4d725d4b7814483f2c840432779c8" gracePeriod=10 Nov 24 13:34:26 crc kubenswrapper[4824]: I1124 13:34:26.565187 4824 generic.go:334] "Generic (PLEG): container finished" podID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" containerID="8a39ea221e0a460c01cbe5046e500c3a42d4d725d4b7814483f2c840432779c8" exitCode=0 Nov 24 13:34:26 crc kubenswrapper[4824]: I1124 13:34:26.565238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" event={"ID":"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc","Type":"ContainerDied","Data":"8a39ea221e0a460c01cbe5046e500c3a42d4d725d4b7814483f2c840432779c8"} Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.444683 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-trpzx" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.452916 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.462219 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8wqh7" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.548054 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmccx\" (UniqueName: \"kubernetes.io/projected/0781aef7-bf19-478e-931f-a10a3dcd586d-kube-api-access-dmccx\") pod \"0781aef7-bf19-478e-931f-a10a3dcd586d\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.548229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-combined-ca-bundle\") pod \"0781aef7-bf19-478e-931f-a10a3dcd586d\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.548285 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-db-sync-config-data\") pod \"0781aef7-bf19-478e-931f-a10a3dcd586d\" (UID: \"0781aef7-bf19-478e-931f-a10a3dcd586d\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.583488 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0781aef7-bf19-478e-931f-a10a3dcd586d" (UID: "0781aef7-bf19-478e-931f-a10a3dcd586d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.603356 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0781aef7-bf19-478e-931f-a10a3dcd586d-kube-api-access-dmccx" (OuterVolumeSpecName: "kube-api-access-dmccx") pod "0781aef7-bf19-478e-931f-a10a3dcd586d" (UID: "0781aef7-bf19-478e-931f-a10a3dcd586d"). InnerVolumeSpecName "kube-api-access-dmccx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.608990 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0781aef7-bf19-478e-931f-a10a3dcd586d" (UID: "0781aef7-bf19-478e-931f-a10a3dcd586d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.613033 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-trpzx" event={"ID":"0781aef7-bf19-478e-931f-a10a3dcd586d","Type":"ContainerDied","Data":"8e7202915d7cd0d11d7d5f6752bf8cf623f1c2956ad82d39fd28e623d2643aca"} Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.613053 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-trpzx" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.613075 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e7202915d7cd0d11d7d5f6752bf8cf623f1c2956ad82d39fd28e623d2643aca" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.624445 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8wqh7" event={"ID":"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce","Type":"ContainerDied","Data":"cec8ea16c244afd03a966f0e00c5a5a7d272e17cafe0555835df2ea30d2ba54e"} Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.624479 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cec8ea16c244afd03a966f0e00c5a5a7d272e17cafe0555835df2ea30d2ba54e" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.624677 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8wqh7" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.626411 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.626507 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.627150 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nhzcv" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.627408 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nhzcv" event={"ID":"4735a096-fa3b-4a42-8ff1-fb1f92b2e953","Type":"ContainerDied","Data":"3fe2986614d7e2a51bfe3d7bb6bb93830334243ef31af9c9ff86492280677e4b"} Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.627485 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fe2986614d7e2a51bfe3d7bb6bb93830334243ef31af9c9ff86492280677e4b" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.627581 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.627641 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.652500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xswzr\" (UniqueName: \"kubernetes.io/projected/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-kube-api-access-xswzr\") pod \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.653991 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-combined-ca-bundle\") pod \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.654334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-scripts\") pod \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.654496 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrtpd\" (UniqueName: \"kubernetes.io/projected/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-kube-api-access-mrtpd\") pod \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.655418 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-config-data\") pod \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.655532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-fernet-keys\") pod \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.655625 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-combined-ca-bundle\") pod \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.655720 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-scripts\") pod \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.655820 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-logs\") pod \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\" (UID: \"aef6bc1b-dfa9-4c16-9940-420f9a7a43ce\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.655907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-config-data\") pod \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.656037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-credential-keys\") pod \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\" (UID: \"4735a096-fa3b-4a42-8ff1-fb1f92b2e953\") " Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.656592 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.656703 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0781aef7-bf19-478e-931f-a10a3dcd586d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.656791 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmccx\" (UniqueName: \"kubernetes.io/projected/0781aef7-bf19-478e-931f-a10a3dcd586d-kube-api-access-dmccx\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.657896 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-logs" (OuterVolumeSpecName: "logs") pod "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" (UID: "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.675254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-kube-api-access-mrtpd" (OuterVolumeSpecName: "kube-api-access-mrtpd") pod "4735a096-fa3b-4a42-8ff1-fb1f92b2e953" (UID: "4735a096-fa3b-4a42-8ff1-fb1f92b2e953"). InnerVolumeSpecName "kube-api-access-mrtpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.683295 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-kube-api-access-xswzr" (OuterVolumeSpecName: "kube-api-access-xswzr") pod "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" (UID: "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce"). InnerVolumeSpecName "kube-api-access-xswzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.688508 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-scripts" (OuterVolumeSpecName: "scripts") pod "4735a096-fa3b-4a42-8ff1-fb1f92b2e953" (UID: "4735a096-fa3b-4a42-8ff1-fb1f92b2e953"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.693070 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4735a096-fa3b-4a42-8ff1-fb1f92b2e953" (UID: "4735a096-fa3b-4a42-8ff1-fb1f92b2e953"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.693206 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4735a096-fa3b-4a42-8ff1-fb1f92b2e953" (UID: "4735a096-fa3b-4a42-8ff1-fb1f92b2e953"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.711757 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-scripts" (OuterVolumeSpecName: "scripts") pod "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" (UID: "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.736338 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-57d8799bcb-x76nj"] Nov 24 13:34:27 crc kubenswrapper[4824]: E1124 13:34:27.738704 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4735a096-fa3b-4a42-8ff1-fb1f92b2e953" containerName="keystone-bootstrap" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.738890 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4735a096-fa3b-4a42-8ff1-fb1f92b2e953" containerName="keystone-bootstrap" Nov 24 13:34:27 crc kubenswrapper[4824]: E1124 13:34:27.738910 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0781aef7-bf19-478e-931f-a10a3dcd586d" containerName="barbican-db-sync" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.738917 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0781aef7-bf19-478e-931f-a10a3dcd586d" containerName="barbican-db-sync" Nov 24 13:34:27 crc kubenswrapper[4824]: E1124 13:34:27.738930 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" containerName="placement-db-sync" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.738935 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" containerName="placement-db-sync" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.739281 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" containerName="placement-db-sync" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.739299 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0781aef7-bf19-478e-931f-a10a3dcd586d" containerName="barbican-db-sync" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.739316 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4735a096-fa3b-4a42-8ff1-fb1f92b2e953" containerName="keystone-bootstrap" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.740011 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.744163 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.744418 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.750045 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4735a096-fa3b-4a42-8ff1-fb1f92b2e953" (UID: "4735a096-fa3b-4a42-8ff1-fb1f92b2e953"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.759366 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xswzr\" (UniqueName: \"kubernetes.io/projected/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-kube-api-access-xswzr\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.759711 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.759820 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.759995 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrtpd\" (UniqueName: \"kubernetes.io/projected/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-kube-api-access-mrtpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.768495 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-config-data" (OuterVolumeSpecName: "config-data") pod "4735a096-fa3b-4a42-8ff1-fb1f92b2e953" (UID: "4735a096-fa3b-4a42-8ff1-fb1f92b2e953"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.782338 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.782365 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.782408 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.782428 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.789788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" (UID: "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.800758 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-57d8799bcb-x76nj"] Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.836372 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-config-data" (OuterVolumeSpecName: "config-data") pod "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" (UID: "aef6bc1b-dfa9-4c16-9940-420f9a7a43ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.884745 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-fernet-keys\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.899358 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-config-data\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.905011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-combined-ca-bundle\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.905127 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-internal-tls-certs\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.905241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-scripts\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.905327 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-public-tls-certs\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.905373 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-credential-keys\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.905429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2v6s\" (UniqueName: \"kubernetes.io/projected/378c23b7-2f70-47a0-b605-67087ed59e70-kube-api-access-n2v6s\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.905584 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.905603 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.905616 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4735a096-fa3b-4a42-8ff1-fb1f92b2e953-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:27 crc kubenswrapper[4824]: I1124 13:34:27.957025 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.012955 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-scripts\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.013008 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-public-tls-certs\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.013036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-credential-keys\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.013075 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2v6s\" (UniqueName: \"kubernetes.io/projected/378c23b7-2f70-47a0-b605-67087ed59e70-kube-api-access-n2v6s\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.013194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-fernet-keys\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.013228 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-config-data\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.013255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-combined-ca-bundle\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.013307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-internal-tls-certs\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.026998 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-fernet-keys\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.028821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-config-data\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.029029 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-scripts\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.029508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-internal-tls-certs\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.031443 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-credential-keys\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.037439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-combined-ca-bundle\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.045674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/378c23b7-2f70-47a0-b605-67087ed59e70-public-tls-certs\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.045704 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2v6s\" (UniqueName: \"kubernetes.io/projected/378c23b7-2f70-47a0-b605-67087ed59e70-kube-api-access-n2v6s\") pod \"keystone-57d8799bcb-x76nj\" (UID: \"378c23b7-2f70-47a0-b605-67087ed59e70\") " pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.082278 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.114766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-nb\") pod \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.114974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-config\") pod \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.115062 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-svc\") pod \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.115163 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh6dp\" (UniqueName: \"kubernetes.io/projected/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-kube-api-access-dh6dp\") pod \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.115250 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-swift-storage-0\") pod \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.115328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-sb\") pod \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\" (UID: \"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc\") " Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.125959 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-kube-api-access-dh6dp" (OuterVolumeSpecName: "kube-api-access-dh6dp") pod "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" (UID: "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc"). InnerVolumeSpecName "kube-api-access-dh6dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.216749 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" (UID: "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.220532 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.220557 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh6dp\" (UniqueName: \"kubernetes.io/projected/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-kube-api-access-dh6dp\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.234491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" (UID: "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.238308 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" (UID: "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.245462 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" (UID: "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.281349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-config" (OuterVolumeSpecName: "config") pod "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" (UID: "5a41142f-fdce-44f4-9e1b-4e7ad33d9efc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.323072 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.323103 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.323113 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.323124 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.378386 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-57d8799bcb-x76nj"] Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.631377 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bdb7b6678-pnfdp"] Nov 24 13:34:28 crc kubenswrapper[4824]: E1124 13:34:28.631773 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" containerName="init" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.631789 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" containerName="init" Nov 24 13:34:28 crc kubenswrapper[4824]: E1124 13:34:28.631861 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" containerName="dnsmasq-dns" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.631872 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" containerName="dnsmasq-dns" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.632059 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" containerName="dnsmasq-dns" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.635541 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.638675 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.638894 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pgbzs" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.638962 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.639032 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.647857 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.655135 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bdb7b6678-pnfdp"] Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.658448 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-57d8799bcb-x76nj" event={"ID":"378c23b7-2f70-47a0-b605-67087ed59e70","Type":"ContainerStarted","Data":"7a29303d5740727ba36959a6dd9f995fef9c457781b09df066fe256c1460364f"} Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.661742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"597521ee-8310-4c63-bdd2-aeef878fb8b9","Type":"ContainerStarted","Data":"88184ee23cb0537c7092fb7b3a1cc6fb588814996ba5c35f00ac9367b38df909"} Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.663610 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" event={"ID":"5a41142f-fdce-44f4-9e1b-4e7ad33d9efc","Type":"ContainerDied","Data":"bd0d8b4c1719af343a39a7ecb0d7947eff371b58bd181bd3c09e54b3350ad8f7"} Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.663653 4824 scope.go:117] "RemoveContainer" containerID="8a39ea221e0a460c01cbe5046e500c3a42d4d725d4b7814483f2c840432779c8" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.663824 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.729013 4824 scope.go:117] "RemoveContainer" containerID="7012cee1033aa5aa4ff422eeed785e1227e70a31c8fcfda4dd31833432ac72a4" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.732828 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-config-data\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.733053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-scripts\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.733100 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-internal-tls-certs\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.733165 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-public-tls-certs\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.733212 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-logs\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.733431 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp96c\" (UniqueName: \"kubernetes.io/projected/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-kube-api-access-pp96c\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.733457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-combined-ca-bundle\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.841716 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-ktmj4"] Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.844910 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp96c\" (UniqueName: \"kubernetes.io/projected/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-kube-api-access-pp96c\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.844970 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-combined-ca-bundle\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.845062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-config-data\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.845140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-scripts\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.845160 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-internal-tls-certs\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.845205 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-public-tls-certs\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.845231 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-logs\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.845738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-logs\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.853010 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-internal-tls-certs\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.862602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-config-data\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.863598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-public-tls-certs\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.865115 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-scripts\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.865463 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-combined-ca-bundle\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.884406 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-ktmj4"] Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.898698 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6c785d8464-nfqng"] Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.906507 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.910156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp96c\" (UniqueName: \"kubernetes.io/projected/d2b3f41b-4bd4-4fcf-b248-6efd3b68f512-kube-api-access-pp96c\") pod \"placement-bdb7b6678-pnfdp\" (UID: \"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512\") " pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.914992 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-vd664" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.915219 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.915321 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.949686 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6c785d8464-nfqng"] Nov 24 13:34:28 crc kubenswrapper[4824]: I1124 13:34:28.973393 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.000881 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5b664bf49f-czcvh"] Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.002512 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.023739 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057631 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6534eeea-5982-4010-a073-f307d7d58036-config-data-custom\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b227cafd-bcc4-4e74-96ea-0935efd47fcc-logs\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057726 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b227cafd-bcc4-4e74-96ea-0935efd47fcc-config-data\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057794 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtmzf\" (UniqueName: \"kubernetes.io/projected/b227cafd-bcc4-4e74-96ea-0935efd47fcc-kube-api-access-xtmzf\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b227cafd-bcc4-4e74-96ea-0935efd47fcc-combined-ca-bundle\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057882 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk7n7\" (UniqueName: \"kubernetes.io/projected/6534eeea-5982-4010-a073-f307d7d58036-kube-api-access-hk7n7\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057904 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6534eeea-5982-4010-a073-f307d7d58036-logs\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b227cafd-bcc4-4e74-96ea-0935efd47fcc-config-data-custom\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057938 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6534eeea-5982-4010-a073-f307d7d58036-combined-ca-bundle\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.057955 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6534eeea-5982-4010-a073-f307d7d58036-config-data\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.073042 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" path="/var/lib/kubelet/pods/5a41142f-fdce-44f4-9e1b-4e7ad33d9efc/volumes" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.074029 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b664bf49f-czcvh"] Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.074126 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-lcgn7"] Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.075449 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.083858 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-lcgn7"] Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.095864 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c8bc7b4c4-w6hnk"] Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.100595 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.106152 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.106664 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c8bc7b4c4-w6hnk"] Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.158989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6534eeea-5982-4010-a073-f307d7d58036-logs\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-combined-ca-bundle\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159055 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2z9p\" (UniqueName: \"kubernetes.io/projected/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-kube-api-access-t2z9p\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b227cafd-bcc4-4e74-96ea-0935efd47fcc-config-data-custom\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159097 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159115 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6534eeea-5982-4010-a073-f307d7d58036-combined-ca-bundle\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159134 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6534eeea-5982-4010-a073-f307d7d58036-config-data\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159169 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6534eeea-5982-4010-a073-f307d7d58036-config-data-custom\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159184 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssztq\" (UniqueName: \"kubernetes.io/projected/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-kube-api-access-ssztq\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159202 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b227cafd-bcc4-4e74-96ea-0935efd47fcc-logs\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159219 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-config\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159264 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-svc\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159311 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b227cafd-bcc4-4e74-96ea-0935efd47fcc-config-data\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159327 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-logs\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159397 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159428 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data-custom\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159471 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtmzf\" (UniqueName: \"kubernetes.io/projected/b227cafd-bcc4-4e74-96ea-0935efd47fcc-kube-api-access-xtmzf\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159508 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b227cafd-bcc4-4e74-96ea-0935efd47fcc-combined-ca-bundle\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.159528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk7n7\" (UniqueName: \"kubernetes.io/projected/6534eeea-5982-4010-a073-f307d7d58036-kube-api-access-hk7n7\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.160353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6534eeea-5982-4010-a073-f307d7d58036-logs\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.172520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6534eeea-5982-4010-a073-f307d7d58036-config-data-custom\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.177121 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b227cafd-bcc4-4e74-96ea-0935efd47fcc-logs\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.179329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b227cafd-bcc4-4e74-96ea-0935efd47fcc-config-data-custom\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.208082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtmzf\" (UniqueName: \"kubernetes.io/projected/b227cafd-bcc4-4e74-96ea-0935efd47fcc-kube-api-access-xtmzf\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.215724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b227cafd-bcc4-4e74-96ea-0935efd47fcc-config-data\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.221749 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk7n7\" (UniqueName: \"kubernetes.io/projected/6534eeea-5982-4010-a073-f307d7d58036-kube-api-access-hk7n7\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.221993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b227cafd-bcc4-4e74-96ea-0935efd47fcc-combined-ca-bundle\") pod \"barbican-worker-5b664bf49f-czcvh\" (UID: \"b227cafd-bcc4-4e74-96ea-0935efd47fcc\") " pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.222397 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6534eeea-5982-4010-a073-f307d7d58036-combined-ca-bundle\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.234624 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6534eeea-5982-4010-a073-f307d7d58036-config-data\") pod \"barbican-keystone-listener-6c785d8464-nfqng\" (UID: \"6534eeea-5982-4010-a073-f307d7d58036\") " pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.254194 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.266915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssztq\" (UniqueName: \"kubernetes.io/projected/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-kube-api-access-ssztq\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267047 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-config\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267134 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-svc\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267264 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-logs\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267452 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data-custom\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267533 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267606 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267688 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-combined-ca-bundle\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2z9p\" (UniqueName: \"kubernetes.io/projected/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-kube-api-access-t2z9p\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.267847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.268699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.269491 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-config\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.270096 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-svc\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.270388 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-logs\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.273963 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.280901 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.281514 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.285560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-combined-ca-bundle\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.286183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data-custom\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.290416 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssztq\" (UniqueName: \"kubernetes.io/projected/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-kube-api-access-ssztq\") pod \"barbican-api-6c8bc7b4c4-w6hnk\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.321509 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2z9p\" (UniqueName: \"kubernetes.io/projected/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-kube-api-access-t2z9p\") pod \"dnsmasq-dns-85ff748b95-lcgn7\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.354271 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b664bf49f-czcvh" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.418928 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.421417 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bdb7b6678-pnfdp"] Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.451269 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.466905 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.467401 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.498123 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 24 13:34:29 crc kubenswrapper[4824]: W1124 13:34:29.526390 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2b3f41b_4bd4_4fcf_b248_6efd3b68f512.slice/crio-0869d4a35587e2e0bd5f3d2d98fb9ab52f6bce2bebdeae7c65fded8d7c1473c0 WatchSource:0}: Error finding container 0869d4a35587e2e0bd5f3d2d98fb9ab52f6bce2bebdeae7c65fded8d7c1473c0: Status 404 returned error can't find the container with id 0869d4a35587e2e0bd5f3d2d98fb9ab52f6bce2bebdeae7c65fded8d7c1473c0 Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.697553 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-959bb6dd8-8pcrj" podUID="5828a51f-d146-44ea-b8ea-423e017939d8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.759799 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-57d8799bcb-x76nj" event={"ID":"378c23b7-2f70-47a0-b605-67087ed59e70","Type":"ContainerStarted","Data":"0e9d65f0c824bb673234f525636de7268ac1d075ecd5fc6670deacdea024e15f"} Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.760548 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.761665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bdb7b6678-pnfdp" event={"ID":"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512","Type":"ContainerStarted","Data":"0869d4a35587e2e0bd5f3d2d98fb9ab52f6bce2bebdeae7c65fded8d7c1473c0"} Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.795156 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-57d8799bcb-x76nj" podStartSLOduration=2.79513595 podStartE2EDuration="2.79513595s" podCreationTimestamp="2025-11-24 13:34:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:29.785251495 +0000 UTC m=+1211.424790805" watchObservedRunningTime="2025-11-24 13:34:29.79513595 +0000 UTC m=+1211.434675260" Nov 24 13:34:29 crc kubenswrapper[4824]: I1124 13:34:29.847930 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6c785d8464-nfqng"] Nov 24 13:34:29 crc kubenswrapper[4824]: W1124 13:34:29.865925 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6534eeea_5982_4010_a073_f307d7d58036.slice/crio-eea93d55fb499968c41a855ba6252816d04efb178c516008ca1171aa1114e7eb WatchSource:0}: Error finding container eea93d55fb499968c41a855ba6252816d04efb178c516008ca1171aa1114e7eb: Status 404 returned error can't find the container with id eea93d55fb499968c41a855ba6252816d04efb178c516008ca1171aa1114e7eb Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.124253 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-lcgn7"] Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.209815 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b664bf49f-czcvh"] Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.358297 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c8bc7b4c4-w6hnk"] Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.786272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" event={"ID":"6534eeea-5982-4010-a073-f307d7d58036","Type":"ContainerStarted","Data":"eea93d55fb499968c41a855ba6252816d04efb178c516008ca1171aa1114e7eb"} Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.797034 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b664bf49f-czcvh" event={"ID":"b227cafd-bcc4-4e74-96ea-0935efd47fcc","Type":"ContainerStarted","Data":"1e099339e1880cf43cc646c006dfdeafea77c6f2d911d790834d18cfded081fe"} Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.819125 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bdb7b6678-pnfdp" event={"ID":"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512","Type":"ContainerStarted","Data":"d49de296ff1d43d2077902a475ec8dc215185159da08e3e2e220c0595df2aa30"} Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.819396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bdb7b6678-pnfdp" event={"ID":"d2b3f41b-4bd4-4fcf-b248-6efd3b68f512","Type":"ContainerStarted","Data":"7c249a940588bf22cf2a2bbd70bb1b66df05f1d5b02f723df2388b947923fbaa"} Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.819440 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.819463 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.824068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" event={"ID":"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f","Type":"ContainerStarted","Data":"a14081e678931eea8fdf708c37db8daf0451f641a9c5c4a375958b624565b352"} Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.829469 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" event={"ID":"6e8c3d20-1c36-4bbe-9176-3bfb989702c3","Type":"ContainerStarted","Data":"fe1abd3bba1d5aa7974ad7372061220c105d76c55549ba3e8c92376c1024818b"} Nov 24 13:34:30 crc kubenswrapper[4824]: I1124 13:34:30.845294 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-bdb7b6678-pnfdp" podStartSLOduration=2.845273632 podStartE2EDuration="2.845273632s" podCreationTimestamp="2025-11-24 13:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:30.840199311 +0000 UTC m=+1212.479738621" watchObservedRunningTime="2025-11-24 13:34:30.845273632 +0000 UTC m=+1212.484812942" Nov 24 13:34:31 crc kubenswrapper[4824]: I1124 13:34:31.846374 4824 generic.go:334] "Generic (PLEG): container finished" podID="6e8c3d20-1c36-4bbe-9176-3bfb989702c3" containerID="7cf598350f47422e877c652160c24677cce39a786288ffb70dc49ca4b5475592" exitCode=0 Nov 24 13:34:31 crc kubenswrapper[4824]: I1124 13:34:31.846623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" event={"ID":"6e8c3d20-1c36-4bbe-9176-3bfb989702c3","Type":"ContainerDied","Data":"7cf598350f47422e877c652160c24677cce39a786288ffb70dc49ca4b5475592"} Nov 24 13:34:31 crc kubenswrapper[4824]: I1124 13:34:31.850863 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-69zcz" event={"ID":"b1af3874-3566-4785-a5fe-13375aa658c3","Type":"ContainerStarted","Data":"e05c7e2a822f0ecac3863cc30b8a66eea207ed911e30a1572205c34f80baddfa"} Nov 24 13:34:31 crc kubenswrapper[4824]: I1124 13:34:31.898892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" event={"ID":"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f","Type":"ContainerStarted","Data":"5cfe786cd566b4a539062963d92056c0917f82c46897138adcb18370f294cfc2"} Nov 24 13:34:31 crc kubenswrapper[4824]: I1124 13:34:31.898939 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" event={"ID":"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f","Type":"ContainerStarted","Data":"af7b85cdb856695836531bde8c6183612584f64259962d5764ae5c7e8e2af01d"} Nov 24 13:34:31 crc kubenswrapper[4824]: I1124 13:34:31.898972 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:31 crc kubenswrapper[4824]: I1124 13:34:31.899091 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:31 crc kubenswrapper[4824]: I1124 13:34:31.943883 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" podStartSLOduration=3.9438705499999998 podStartE2EDuration="3.94387055s" podCreationTimestamp="2025-11-24 13:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:31.940646197 +0000 UTC m=+1213.580185507" watchObservedRunningTime="2025-11-24 13:34:31.94387055 +0000 UTC m=+1213.583409850" Nov 24 13:34:31 crc kubenswrapper[4824]: I1124 13:34:31.965179 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-69zcz" podStartSLOduration=3.345455194 podStartE2EDuration="56.965161039s" podCreationTimestamp="2025-11-24 13:33:35 +0000 UTC" firstStartedPulling="2025-11-24 13:33:36.952278989 +0000 UTC m=+1158.591818299" lastFinishedPulling="2025-11-24 13:34:30.571984834 +0000 UTC m=+1212.211524144" observedRunningTime="2025-11-24 13:34:31.964459141 +0000 UTC m=+1213.603998451" watchObservedRunningTime="2025-11-24 13:34:31.965161039 +0000 UTC m=+1213.604700349" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.684172 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7545588bcd-5szvs"] Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.686002 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.688980 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.689171 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.717160 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7545588bcd-5szvs"] Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.770818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-logs\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.770870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-config-data\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.770920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-public-tls-certs\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.770941 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-config-data-custom\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.770955 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-combined-ca-bundle\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.770977 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4sq5\" (UniqueName: \"kubernetes.io/projected/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-kube-api-access-d4sq5\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.771047 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-internal-tls-certs\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.829456 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-ktmj4" podUID="5a41142f-fdce-44f4-9e1b-4e7ad33d9efc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.142:5353: i/o timeout" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.872478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-internal-tls-certs\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.872598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-logs\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.872626 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-config-data\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.873732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-logs\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.874068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-public-tls-certs\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.874095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-config-data-custom\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.874113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-combined-ca-bundle\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.874132 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4sq5\" (UniqueName: \"kubernetes.io/projected/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-kube-api-access-d4sq5\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.884989 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-combined-ca-bundle\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.895855 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-internal-tls-certs\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.900031 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-public-tls-certs\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.900698 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-config-data-custom\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.913240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4sq5\" (UniqueName: \"kubernetes.io/projected/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-kube-api-access-d4sq5\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.914218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960-config-data\") pod \"barbican-api-7545588bcd-5szvs\" (UID: \"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960\") " pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.949368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" event={"ID":"6e8c3d20-1c36-4bbe-9176-3bfb989702c3","Type":"ContainerStarted","Data":"7e41b82340ab499c13c31db849ba190fc51479bd1c5f1dd1b5676f412b31425d"} Nov 24 13:34:32 crc kubenswrapper[4824]: I1124 13:34:32.949443 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:33 crc kubenswrapper[4824]: I1124 13:34:33.018261 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:33 crc kubenswrapper[4824]: I1124 13:34:33.049779 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" podStartSLOduration=5.049762398 podStartE2EDuration="5.049762398s" podCreationTimestamp="2025-11-24 13:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:32.976187253 +0000 UTC m=+1214.615726583" watchObservedRunningTime="2025-11-24 13:34:33.049762398 +0000 UTC m=+1214.689301708" Nov 24 13:34:33 crc kubenswrapper[4824]: I1124 13:34:33.072193 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 13:34:33 crc kubenswrapper[4824]: I1124 13:34:33.072848 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 13:34:33 crc kubenswrapper[4824]: I1124 13:34:33.093218 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:33 crc kubenswrapper[4824]: I1124 13:34:33.093327 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:34:33 crc kubenswrapper[4824]: I1124 13:34:33.224876 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 13:34:36 crc kubenswrapper[4824]: I1124 13:34:36.012247 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" event={"ID":"6534eeea-5982-4010-a073-f307d7d58036","Type":"ContainerStarted","Data":"8af7082a6d5cf4f7cdb8aff00bb5818d862162cf11f8ddcfc0714b3a4e679f55"} Nov 24 13:34:36 crc kubenswrapper[4824]: I1124 13:34:36.029207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b664bf49f-czcvh" event={"ID":"b227cafd-bcc4-4e74-96ea-0935efd47fcc","Type":"ContainerStarted","Data":"9b5209e9eea5d491e06975f8bfc6dad2f245957793206c95bcac4f3e58ad3648"} Nov 24 13:34:36 crc kubenswrapper[4824]: I1124 13:34:36.033577 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7545588bcd-5szvs"] Nov 24 13:34:36 crc kubenswrapper[4824]: W1124 13:34:36.051915 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd19d5c0e_cf0b_46a8_b1b7_e832c2fd2960.slice/crio-2671714fdec0a9ef63fc289ecc038a82608f467e2e0e6c22c958bd2ad120c5e5 WatchSource:0}: Error finding container 2671714fdec0a9ef63fc289ecc038a82608f467e2e0e6c22c958bd2ad120c5e5: Status 404 returned error can't find the container with id 2671714fdec0a9ef63fc289ecc038a82608f467e2e0e6c22c958bd2ad120c5e5 Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.050445 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" event={"ID":"6534eeea-5982-4010-a073-f307d7d58036","Type":"ContainerStarted","Data":"fa53d3898dc6c6cf1a2e4d87b16cbab85ec6b6359f9440bce6b3e62d5c520eff"} Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.062997 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b664bf49f-czcvh" event={"ID":"b227cafd-bcc4-4e74-96ea-0935efd47fcc","Type":"ContainerStarted","Data":"96782fbaee178cc3fa8b9d0e9e0b2acaac1ec6a299050aa56334d37fbd6be785"} Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.075180 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6c785d8464-nfqng" podStartSLOduration=3.60053222 podStartE2EDuration="9.075164984s" podCreationTimestamp="2025-11-24 13:34:28 +0000 UTC" firstStartedPulling="2025-11-24 13:34:29.877154532 +0000 UTC m=+1211.516693842" lastFinishedPulling="2025-11-24 13:34:35.351787296 +0000 UTC m=+1216.991326606" observedRunningTime="2025-11-24 13:34:37.067315372 +0000 UTC m=+1218.706854682" watchObservedRunningTime="2025-11-24 13:34:37.075164984 +0000 UTC m=+1218.714704294" Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.076521 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7545588bcd-5szvs" event={"ID":"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960","Type":"ContainerStarted","Data":"704ec4b05d568b372492b5339546e8dda49f215ac55cbf4bd104397c05cfb6a3"} Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.076562 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7545588bcd-5szvs" event={"ID":"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960","Type":"ContainerStarted","Data":"0affd8333c11c223023e2b5faa7ef8adcb614fc6b5c8b2740f9f061510e33c2e"} Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.076574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7545588bcd-5szvs" event={"ID":"d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960","Type":"ContainerStarted","Data":"2671714fdec0a9ef63fc289ecc038a82608f467e2e0e6c22c958bd2ad120c5e5"} Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.076734 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.076846 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.103398 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5b664bf49f-czcvh" podStartSLOduration=3.9889921040000003 podStartE2EDuration="9.103377181s" podCreationTimestamp="2025-11-24 13:34:28 +0000 UTC" firstStartedPulling="2025-11-24 13:34:30.237912842 +0000 UTC m=+1211.877452152" lastFinishedPulling="2025-11-24 13:34:35.352297919 +0000 UTC m=+1216.991837229" observedRunningTime="2025-11-24 13:34:37.09945069 +0000 UTC m=+1218.738989990" watchObservedRunningTime="2025-11-24 13:34:37.103377181 +0000 UTC m=+1218.742916481" Nov 24 13:34:37 crc kubenswrapper[4824]: I1124 13:34:37.118938 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7545588bcd-5szvs" podStartSLOduration=5.118918261 podStartE2EDuration="5.118918261s" podCreationTimestamp="2025-11-24 13:34:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:37.116514809 +0000 UTC m=+1218.756054119" watchObservedRunningTime="2025-11-24 13:34:37.118918261 +0000 UTC m=+1218.758457571" Nov 24 13:34:39 crc kubenswrapper[4824]: I1124 13:34:39.421534 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:39 crc kubenswrapper[4824]: I1124 13:34:39.469384 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 24 13:34:39 crc kubenswrapper[4824]: I1124 13:34:39.488096 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r9wlh"] Nov 24 13:34:39 crc kubenswrapper[4824]: I1124 13:34:39.488386 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" podUID="984c9282-012e-443d-b80a-89133d02dcde" containerName="dnsmasq-dns" containerID="cri-o://58e8551b4a63b882692d2d8d91f967797db29235af2bf39d0874dfd9fd229779" gracePeriod=10 Nov 24 13:34:39 crc kubenswrapper[4824]: I1124 13:34:39.678417 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-959bb6dd8-8pcrj" podUID="5828a51f-d146-44ea-b8ea-423e017939d8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 24 13:34:40 crc kubenswrapper[4824]: I1124 13:34:40.114228 4824 generic.go:334] "Generic (PLEG): container finished" podID="984c9282-012e-443d-b80a-89133d02dcde" containerID="58e8551b4a63b882692d2d8d91f967797db29235af2bf39d0874dfd9fd229779" exitCode=0 Nov 24 13:34:40 crc kubenswrapper[4824]: I1124 13:34:40.114265 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" event={"ID":"984c9282-012e-443d-b80a-89133d02dcde","Type":"ContainerDied","Data":"58e8551b4a63b882692d2d8d91f967797db29235af2bf39d0874dfd9fd229779"} Nov 24 13:34:40 crc kubenswrapper[4824]: I1124 13:34:40.787911 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:34:40 crc kubenswrapper[4824]: I1124 13:34:40.788221 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:34:40 crc kubenswrapper[4824]: I1124 13:34:40.916550 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:34:41 crc kubenswrapper[4824]: I1124 13:34:41.084618 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" podUID="984c9282-012e-443d-b80a-89133d02dcde" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.152:5353: connect: connection refused" Nov 24 13:34:41 crc kubenswrapper[4824]: I1124 13:34:41.319532 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:41 crc kubenswrapper[4824]: I1124 13:34:41.438413 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:43 crc kubenswrapper[4824]: I1124 13:34:43.165093 4824 generic.go:334] "Generic (PLEG): container finished" podID="b1af3874-3566-4785-a5fe-13375aa658c3" containerID="e05c7e2a822f0ecac3863cc30b8a66eea207ed911e30a1572205c34f80baddfa" exitCode=0 Nov 24 13:34:43 crc kubenswrapper[4824]: I1124 13:34:43.165176 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-69zcz" event={"ID":"b1af3874-3566-4785-a5fe-13375aa658c3","Type":"ContainerDied","Data":"e05c7e2a822f0ecac3863cc30b8a66eea207ed911e30a1572205c34f80baddfa"} Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.735758 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.759210 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-69zcz" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.765216 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833196 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-config-data\") pod \"b1af3874-3566-4785-a5fe-13375aa658c3\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833310 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-combined-ca-bundle\") pod \"b1af3874-3566-4785-a5fe-13375aa658c3\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833339 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dpm7\" (UniqueName: \"kubernetes.io/projected/b1af3874-3566-4785-a5fe-13375aa658c3-kube-api-access-5dpm7\") pod \"b1af3874-3566-4785-a5fe-13375aa658c3\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833373 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-scripts\") pod \"b1af3874-3566-4785-a5fe-13375aa658c3\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833405 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-svc\") pod \"984c9282-012e-443d-b80a-89133d02dcde\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-swift-storage-0\") pod \"984c9282-012e-443d-b80a-89133d02dcde\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-db-sync-config-data\") pod \"b1af3874-3566-4785-a5fe-13375aa658c3\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833471 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-sb\") pod \"984c9282-012e-443d-b80a-89133d02dcde\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833495 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6rzg\" (UniqueName: \"kubernetes.io/projected/984c9282-012e-443d-b80a-89133d02dcde-kube-api-access-g6rzg\") pod \"984c9282-012e-443d-b80a-89133d02dcde\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833547 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-config\") pod \"984c9282-012e-443d-b80a-89133d02dcde\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833567 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-nb\") pod \"984c9282-012e-443d-b80a-89133d02dcde\" (UID: \"984c9282-012e-443d-b80a-89133d02dcde\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.833611 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1af3874-3566-4785-a5fe-13375aa658c3-etc-machine-id\") pod \"b1af3874-3566-4785-a5fe-13375aa658c3\" (UID: \"b1af3874-3566-4785-a5fe-13375aa658c3\") " Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.834967 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1af3874-3566-4785-a5fe-13375aa658c3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b1af3874-3566-4785-a5fe-13375aa658c3" (UID: "b1af3874-3566-4785-a5fe-13375aa658c3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.846434 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/984c9282-012e-443d-b80a-89133d02dcde-kube-api-access-g6rzg" (OuterVolumeSpecName: "kube-api-access-g6rzg") pod "984c9282-012e-443d-b80a-89133d02dcde" (UID: "984c9282-012e-443d-b80a-89133d02dcde"). InnerVolumeSpecName "kube-api-access-g6rzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.849132 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b1af3874-3566-4785-a5fe-13375aa658c3" (UID: "b1af3874-3566-4785-a5fe-13375aa658c3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.875213 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-scripts" (OuterVolumeSpecName: "scripts") pod "b1af3874-3566-4785-a5fe-13375aa658c3" (UID: "b1af3874-3566-4785-a5fe-13375aa658c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.916288 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1af3874-3566-4785-a5fe-13375aa658c3-kube-api-access-5dpm7" (OuterVolumeSpecName: "kube-api-access-5dpm7") pod "b1af3874-3566-4785-a5fe-13375aa658c3" (UID: "b1af3874-3566-4785-a5fe-13375aa658c3"). InnerVolumeSpecName "kube-api-access-5dpm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.935467 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.935488 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.935499 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6rzg\" (UniqueName: \"kubernetes.io/projected/984c9282-012e-443d-b80a-89133d02dcde-kube-api-access-g6rzg\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.935508 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1af3874-3566-4785-a5fe-13375aa658c3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.935517 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dpm7\" (UniqueName: \"kubernetes.io/projected/b1af3874-3566-4785-a5fe-13375aa658c3-kube-api-access-5dpm7\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.951241 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "984c9282-012e-443d-b80a-89133d02dcde" (UID: "984c9282-012e-443d-b80a-89133d02dcde"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.966410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1af3874-3566-4785-a5fe-13375aa658c3" (UID: "b1af3874-3566-4785-a5fe-13375aa658c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.974396 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-config-data" (OuterVolumeSpecName: "config-data") pod "b1af3874-3566-4785-a5fe-13375aa658c3" (UID: "b1af3874-3566-4785-a5fe-13375aa658c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.994036 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "984c9282-012e-443d-b80a-89133d02dcde" (UID: "984c9282-012e-443d-b80a-89133d02dcde"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:44 crc kubenswrapper[4824]: I1124 13:34:44.998326 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "984c9282-012e-443d-b80a-89133d02dcde" (UID: "984c9282-012e-443d-b80a-89133d02dcde"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.008643 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-config" (OuterVolumeSpecName: "config") pod "984c9282-012e-443d-b80a-89133d02dcde" (UID: "984c9282-012e-443d-b80a-89133d02dcde"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.014362 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "984c9282-012e-443d-b80a-89133d02dcde" (UID: "984c9282-012e-443d-b80a-89133d02dcde"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.036871 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.037082 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.037144 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.037212 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.037288 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.037345 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/984c9282-012e-443d-b80a-89133d02dcde-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.037403 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1af3874-3566-4785-a5fe-13375aa658c3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:45 crc kubenswrapper[4824]: E1124 13:34:45.062560 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.182341 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-69zcz" event={"ID":"b1af3874-3566-4785-a5fe-13375aa658c3","Type":"ContainerDied","Data":"38d5c778c8d8e21b4fcfffc530c3cd1bad8dd322e0cba7ad045f473dc3875cff"} Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.182631 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38d5c778c8d8e21b4fcfffc530c3cd1bad8dd322e0cba7ad045f473dc3875cff" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.182360 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-69zcz" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.184504 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"597521ee-8310-4c63-bdd2-aeef878fb8b9","Type":"ContainerStarted","Data":"2273b867bb68de04ba414993907aa4c3c1c9148de145ecf5972037b9bc3292ea"} Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.184654 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="ceilometer-notification-agent" containerID="cri-o://a6898c246a7e470e32264232cbde12bde5dfa4a4dd33ccaf1bb34d408b7cbdb6" gracePeriod=30 Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.184912 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="sg-core" containerID="cri-o://88184ee23cb0537c7092fb7b3a1cc6fb588814996ba5c35f00ac9367b38df909" gracePeriod=30 Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.184992 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="proxy-httpd" containerID="cri-o://2273b867bb68de04ba414993907aa4c3c1c9148de145ecf5972037b9bc3292ea" gracePeriod=30 Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.185973 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.195976 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" event={"ID":"984c9282-012e-443d-b80a-89133d02dcde","Type":"ContainerDied","Data":"e896deeeac3e748d459f4b76fa9db8a670ee3ecc3906a06855128fd14dcc9865"} Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.196029 4824 scope.go:117] "RemoveContainer" containerID="58e8551b4a63b882692d2d8d91f967797db29235af2bf39d0874dfd9fd229779" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.196044 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-r9wlh" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.225972 4824 scope.go:117] "RemoveContainer" containerID="5932ef8564ca81c459a1d6855d567023d3cce56ed33f9014d0d06deb77e7fb43" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.257025 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r9wlh"] Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.263923 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-r9wlh"] Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.383165 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7545588bcd-5szvs" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.495385 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:45 crc kubenswrapper[4824]: E1124 13:34:45.504559 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984c9282-012e-443d-b80a-89133d02dcde" containerName="dnsmasq-dns" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.504755 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="984c9282-012e-443d-b80a-89133d02dcde" containerName="dnsmasq-dns" Nov 24 13:34:45 crc kubenswrapper[4824]: E1124 13:34:45.504834 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984c9282-012e-443d-b80a-89133d02dcde" containerName="init" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.504894 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="984c9282-012e-443d-b80a-89133d02dcde" containerName="init" Nov 24 13:34:45 crc kubenswrapper[4824]: E1124 13:34:45.504962 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1af3874-3566-4785-a5fe-13375aa658c3" containerName="cinder-db-sync" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.505022 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1af3874-3566-4785-a5fe-13375aa658c3" containerName="cinder-db-sync" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.505236 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1af3874-3566-4785-a5fe-13375aa658c3" containerName="cinder-db-sync" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.505302 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="984c9282-012e-443d-b80a-89133d02dcde" containerName="dnsmasq-dns" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.506250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.509227 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.509470 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.509746 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7pc9z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.516423 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.529612 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.601581 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c8bc7b4c4-w6hnk"] Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.601801 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api-log" containerID="cri-o://af7b85cdb856695836531bde8c6183612584f64259962d5764ae5c7e8e2af01d" gracePeriod=30 Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.602200 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api" containerID="cri-o://5cfe786cd566b4a539062963d92056c0917f82c46897138adcb18370f294cfc2" gracePeriod=30 Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.609094 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.648503 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zbz6\" (UniqueName: \"kubernetes.io/projected/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-kube-api-access-2zbz6\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.648542 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.648565 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-scripts\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.648613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.648652 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.648711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.671442 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-k264z"] Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.673949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.697570 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-k264z"] Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zbz6\" (UniqueName: \"kubernetes.io/projected/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-kube-api-access-2zbz6\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751330 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-scripts\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751362 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-config\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751416 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751456 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rhbl\" (UniqueName: \"kubernetes.io/projected/0367b064-fb20-459e-8878-e60a5c8b8207-kube-api-access-2rhbl\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751531 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.751593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.754777 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.760247 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.762193 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.767177 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-scripts\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.786194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.805442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zbz6\" (UniqueName: \"kubernetes.io/projected/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-kube-api-access-2zbz6\") pod \"cinder-scheduler-0\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.842479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.855723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.856533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.856607 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-config\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.857135 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-config\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.857366 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rhbl\" (UniqueName: \"kubernetes.io/projected/0367b064-fb20-459e-8878-e60a5c8b8207-kube-api-access-2rhbl\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.857655 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.857724 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.857754 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.858755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.859159 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.859232 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.905975 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rhbl\" (UniqueName: \"kubernetes.io/projected/0367b064-fb20-459e-8878-e60a5c8b8207-kube-api-access-2rhbl\") pod \"dnsmasq-dns-5c9776ccc5-k264z\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.980069 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.981581 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.990149 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 13:34:45 crc kubenswrapper[4824]: I1124 13:34:45.993011 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.070845 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50ecbaf8-dbbf-4d2c-8394-b0319768e871-etc-machine-id\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.070893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-scripts\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.070913 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50ecbaf8-dbbf-4d2c-8394-b0319768e871-logs\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.071010 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data-custom\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.071041 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.071065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.071080 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd4x9\" (UniqueName: \"kubernetes.io/projected/50ecbaf8-dbbf-4d2c-8394-b0319768e871-kube-api-access-vd4x9\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.162413 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.177956 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50ecbaf8-dbbf-4d2c-8394-b0319768e871-etc-machine-id\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.180400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50ecbaf8-dbbf-4d2c-8394-b0319768e871-etc-machine-id\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.180910 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-scripts\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.180956 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50ecbaf8-dbbf-4d2c-8394-b0319768e871-logs\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.181161 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data-custom\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.181227 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.181278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.181301 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd4x9\" (UniqueName: \"kubernetes.io/projected/50ecbaf8-dbbf-4d2c-8394-b0319768e871-kube-api-access-vd4x9\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.190063 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50ecbaf8-dbbf-4d2c-8394-b0319768e871-logs\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.194797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.202830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data-custom\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.219030 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.219517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-scripts\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.235361 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd4x9\" (UniqueName: \"kubernetes.io/projected/50ecbaf8-dbbf-4d2c-8394-b0319768e871-kube-api-access-vd4x9\") pod \"cinder-api-0\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.263800 4824 generic.go:334] "Generic (PLEG): container finished" podID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerID="af7b85cdb856695836531bde8c6183612584f64259962d5764ae5c7e8e2af01d" exitCode=143 Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.263877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" event={"ID":"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f","Type":"ContainerDied","Data":"af7b85cdb856695836531bde8c6183612584f64259962d5764ae5c7e8e2af01d"} Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.290543 4824 generic.go:334] "Generic (PLEG): container finished" podID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerID="2273b867bb68de04ba414993907aa4c3c1c9148de145ecf5972037b9bc3292ea" exitCode=0 Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.290576 4824 generic.go:334] "Generic (PLEG): container finished" podID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerID="88184ee23cb0537c7092fb7b3a1cc6fb588814996ba5c35f00ac9367b38df909" exitCode=2 Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.290614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"597521ee-8310-4c63-bdd2-aeef878fb8b9","Type":"ContainerDied","Data":"2273b867bb68de04ba414993907aa4c3c1c9148de145ecf5972037b9bc3292ea"} Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.290640 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"597521ee-8310-4c63-bdd2-aeef878fb8b9","Type":"ContainerDied","Data":"88184ee23cb0537c7092fb7b3a1cc6fb588814996ba5c35f00ac9367b38df909"} Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.330213 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.386109 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.595626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.841548 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-k264z"] Nov 24 13:34:46 crc kubenswrapper[4824]: I1124 13:34:46.999120 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:47 crc kubenswrapper[4824]: I1124 13:34:47.058473 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="984c9282-012e-443d-b80a-89133d02dcde" path="/var/lib/kubelet/pods/984c9282-012e-443d-b80a-89133d02dcde/volumes" Nov 24 13:34:47 crc kubenswrapper[4824]: I1124 13:34:47.321944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9","Type":"ContainerStarted","Data":"fc656c368945e6bdebbea039132045aa8c5879a94bf444d3f16c10bfbfca877f"} Nov 24 13:34:47 crc kubenswrapper[4824]: I1124 13:34:47.333856 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50ecbaf8-dbbf-4d2c-8394-b0319768e871","Type":"ContainerStarted","Data":"6ab0f3ebf4e9f0109c7f7e9b5c7261ec4dabc1f1051c9cc9732cd404124cacc2"} Nov 24 13:34:47 crc kubenswrapper[4824]: I1124 13:34:47.337680 4824 generic.go:334] "Generic (PLEG): container finished" podID="0367b064-fb20-459e-8878-e60a5c8b8207" containerID="6bcad28ae4ef545e81c03a052f5b1a3502d9d111cf5add261503dae1bb02340c" exitCode=0 Nov 24 13:34:47 crc kubenswrapper[4824]: I1124 13:34:47.337706 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" event={"ID":"0367b064-fb20-459e-8878-e60a5c8b8207","Type":"ContainerDied","Data":"6bcad28ae4ef545e81c03a052f5b1a3502d9d111cf5add261503dae1bb02340c"} Nov 24 13:34:47 crc kubenswrapper[4824]: I1124 13:34:47.337722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" event={"ID":"0367b064-fb20-459e-8878-e60a5c8b8207","Type":"ContainerStarted","Data":"3b28efe5d0535be381ef997adc0445b7bd1217bb881542e1897516de79c576e2"} Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.336451 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.382324 4824 generic.go:334] "Generic (PLEG): container finished" podID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerID="a6898c246a7e470e32264232cbde12bde5dfa4a4dd33ccaf1bb34d408b7cbdb6" exitCode=0 Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.382409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"597521ee-8310-4c63-bdd2-aeef878fb8b9","Type":"ContainerDied","Data":"a6898c246a7e470e32264232cbde12bde5dfa4a4dd33ccaf1bb34d408b7cbdb6"} Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.407247 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" event={"ID":"0367b064-fb20-459e-8878-e60a5c8b8207","Type":"ContainerStarted","Data":"d477037c13bc616f87cccca7d9817b438eb3db455aea1620a5f6b7a903937b9a"} Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.407401 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.437946 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" podStartSLOduration=3.437927121 podStartE2EDuration="3.437927121s" podCreationTimestamp="2025-11-24 13:34:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:48.431115046 +0000 UTC m=+1230.070654356" watchObservedRunningTime="2025-11-24 13:34:48.437927121 +0000 UTC m=+1230.077466431" Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.812771 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-798f5d656f-r9xcp" Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.857442 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.977328 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d494558c4-pqqd8"] Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.997273 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d494558c4-pqqd8" podUID="a5c8cb45-5223-4339-a957-80ec09aad094" containerName="neutron-api" containerID="cri-o://c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea" gracePeriod=30 Nov 24 13:34:48 crc kubenswrapper[4824]: I1124 13:34:48.997976 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d494558c4-pqqd8" podUID="a5c8cb45-5223-4339-a957-80ec09aad094" containerName="neutron-httpd" containerID="cri-o://f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56" gracePeriod=30 Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.003577 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-log-httpd\") pod \"597521ee-8310-4c63-bdd2-aeef878fb8b9\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.003642 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-config-data\") pod \"597521ee-8310-4c63-bdd2-aeef878fb8b9\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.003665 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-run-httpd\") pod \"597521ee-8310-4c63-bdd2-aeef878fb8b9\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.003757 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-scripts\") pod \"597521ee-8310-4c63-bdd2-aeef878fb8b9\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.013241 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "597521ee-8310-4c63-bdd2-aeef878fb8b9" (UID: "597521ee-8310-4c63-bdd2-aeef878fb8b9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.013541 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "597521ee-8310-4c63-bdd2-aeef878fb8b9" (UID: "597521ee-8310-4c63-bdd2-aeef878fb8b9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.003804 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-sg-core-conf-yaml\") pod \"597521ee-8310-4c63-bdd2-aeef878fb8b9\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.016123 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flfrr\" (UniqueName: \"kubernetes.io/projected/597521ee-8310-4c63-bdd2-aeef878fb8b9-kube-api-access-flfrr\") pod \"597521ee-8310-4c63-bdd2-aeef878fb8b9\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.016193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-combined-ca-bundle\") pod \"597521ee-8310-4c63-bdd2-aeef878fb8b9\" (UID: \"597521ee-8310-4c63-bdd2-aeef878fb8b9\") " Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.017261 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.017282 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/597521ee-8310-4c63-bdd2-aeef878fb8b9-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.073959 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/597521ee-8310-4c63-bdd2-aeef878fb8b9-kube-api-access-flfrr" (OuterVolumeSpecName: "kube-api-access-flfrr") pod "597521ee-8310-4c63-bdd2-aeef878fb8b9" (UID: "597521ee-8310-4c63-bdd2-aeef878fb8b9"). InnerVolumeSpecName "kube-api-access-flfrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.092959 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-scripts" (OuterVolumeSpecName: "scripts") pod "597521ee-8310-4c63-bdd2-aeef878fb8b9" (UID: "597521ee-8310-4c63-bdd2-aeef878fb8b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.120879 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.120908 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flfrr\" (UniqueName: \"kubernetes.io/projected/597521ee-8310-4c63-bdd2-aeef878fb8b9-kube-api-access-flfrr\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.194945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "597521ee-8310-4c63-bdd2-aeef878fb8b9" (UID: "597521ee-8310-4c63-bdd2-aeef878fb8b9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.224113 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.226127 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "597521ee-8310-4c63-bdd2-aeef878fb8b9" (UID: "597521ee-8310-4c63-bdd2-aeef878fb8b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.249972 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-config-data" (OuterVolumeSpecName: "config-data") pod "597521ee-8310-4c63-bdd2-aeef878fb8b9" (UID: "597521ee-8310-4c63-bdd2-aeef878fb8b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.325899 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.325933 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/597521ee-8310-4c63-bdd2-aeef878fb8b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.425111 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"597521ee-8310-4c63-bdd2-aeef878fb8b9","Type":"ContainerDied","Data":"39a54b897d88ef1c661450f89c8643edd5eaed6374f68d36b83b648c539e977b"} Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.425184 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.425386 4824 scope.go:117] "RemoveContainer" containerID="2273b867bb68de04ba414993907aa4c3c1c9148de145ecf5972037b9bc3292ea" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.432851 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9","Type":"ContainerStarted","Data":"47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e"} Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.438044 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50ecbaf8-dbbf-4d2c-8394-b0319768e871","Type":"ContainerStarted","Data":"17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274"} Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.439688 4824 generic.go:334] "Generic (PLEG): container finished" podID="a5c8cb45-5223-4339-a957-80ec09aad094" containerID="f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56" exitCode=0 Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.440529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d494558c4-pqqd8" event={"ID":"a5c8cb45-5223-4339-a957-80ec09aad094","Type":"ContainerDied","Data":"f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56"} Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.466481 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.466547 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.467231 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"6ce765b74f266b883a095df6fbfd4992669f9774df16a85346b8570f4fc1ede0"} pod="openstack/horizon-7c96f4646d-wsbgv" containerMessage="Container horizon failed startup probe, will be restarted" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.467257 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" containerID="cri-o://6ce765b74f266b883a095df6fbfd4992669f9774df16a85346b8570f4fc1ede0" gracePeriod=30 Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.477186 4824 scope.go:117] "RemoveContainer" containerID="88184ee23cb0537c7092fb7b3a1cc6fb588814996ba5c35f00ac9367b38df909" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.511171 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.526462 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.544011 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:49 crc kubenswrapper[4824]: E1124 13:34:49.544347 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="ceilometer-notification-agent" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.544363 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="ceilometer-notification-agent" Nov 24 13:34:49 crc kubenswrapper[4824]: E1124 13:34:49.544389 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="sg-core" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.544395 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="sg-core" Nov 24 13:34:49 crc kubenswrapper[4824]: E1124 13:34:49.544403 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="proxy-httpd" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.544409 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="proxy-httpd" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.544575 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="ceilometer-notification-agent" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.544593 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="sg-core" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.544606 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" containerName="proxy-httpd" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.546209 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.548715 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.548894 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.567394 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.605561 4824 scope.go:117] "RemoveContainer" containerID="a6898c246a7e470e32264232cbde12bde5dfa4a4dd33ccaf1bb34d408b7cbdb6" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.638688 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-log-httpd\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.638764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.638801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh7p4\" (UniqueName: \"kubernetes.io/projected/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-kube-api-access-kh7p4\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.638917 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-config-data\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.639000 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.639053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-run-httpd\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.639161 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-scripts\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.676370 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-959bb6dd8-8pcrj" podUID="5828a51f-d146-44ea-b8ea-423e017939d8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.676454 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.677118 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"f0eb7c3f4f829b0e7b8112bc730b12912e7eb168f9eb8c9618b6934afbce96bd"} pod="openstack/horizon-959bb6dd8-8pcrj" containerMessage="Container horizon failed startup probe, will be restarted" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.677147 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-959bb6dd8-8pcrj" podUID="5828a51f-d146-44ea-b8ea-423e017939d8" containerName="horizon" containerID="cri-o://f0eb7c3f4f829b0e7b8112bc730b12912e7eb168f9eb8c9618b6934afbce96bd" gracePeriod=30 Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.745157 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-run-httpd\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.745206 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-scripts\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.745282 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-log-httpd\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.745332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.745348 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh7p4\" (UniqueName: \"kubernetes.io/projected/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-kube-api-access-kh7p4\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.745375 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-config-data\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.745412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.746405 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-run-httpd\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.750915 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-log-httpd\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.753463 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.754048 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.754102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-scripts\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.755207 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-config-data\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.770526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh7p4\" (UniqueName: \"kubernetes.io/projected/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-kube-api-access-kh7p4\") pod \"ceilometer-0\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " pod="openstack/ceilometer-0" Nov 24 13:34:49 crc kubenswrapper[4824]: I1124 13:34:49.918998 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.380333 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.388331 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7545588bcd-5szvs" podUID="d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.161:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.448160 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerStarted","Data":"85843d22616999c9e1c9eaa8ac48981ed7834b1d4ae28e612b71973855591fed"} Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.454124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9","Type":"ContainerStarted","Data":"62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870"} Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.455345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50ecbaf8-dbbf-4d2c-8394-b0319768e871","Type":"ContainerStarted","Data":"c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff"} Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.455591 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerName="cinder-api-log" containerID="cri-o://17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274" gracePeriod=30 Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.455631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.455657 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerName="cinder-api" containerID="cri-o://c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff" gracePeriod=30 Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.481450 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.79944005 podStartE2EDuration="5.481430202s" podCreationTimestamp="2025-11-24 13:34:45 +0000 UTC" firstStartedPulling="2025-11-24 13:34:46.602532288 +0000 UTC m=+1228.242071608" lastFinishedPulling="2025-11-24 13:34:47.28452245 +0000 UTC m=+1228.924061760" observedRunningTime="2025-11-24 13:34:50.470311516 +0000 UTC m=+1232.109850846" watchObservedRunningTime="2025-11-24 13:34:50.481430202 +0000 UTC m=+1232.120969512" Nov 24 13:34:50 crc kubenswrapper[4824]: I1124 13:34:50.843203 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.020962 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="597521ee-8310-4c63-bdd2-aeef878fb8b9" path="/var/lib/kubelet/pods/597521ee-8310-4c63-bdd2-aeef878fb8b9/volumes" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.244463 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:55146->10.217.0.160:9311: read: connection reset by peer" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.245365 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:55156->10.217.0.160:9311: read: connection reset by peer" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.468431 4824 generic.go:334] "Generic (PLEG): container finished" podID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerID="17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274" exitCode=143 Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.468819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50ecbaf8-dbbf-4d2c-8394-b0319768e871","Type":"ContainerDied","Data":"17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274"} Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.471628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerStarted","Data":"3de161309266940d53a6d44c8aecccb0b7d77ddf2238234eedd6c4003797aa69"} Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.473467 4824 generic.go:334] "Generic (PLEG): container finished" podID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerID="5cfe786cd566b4a539062963d92056c0917f82c46897138adcb18370f294cfc2" exitCode=0 Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.474242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" event={"ID":"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f","Type":"ContainerDied","Data":"5cfe786cd566b4a539062963d92056c0917f82c46897138adcb18370f294cfc2"} Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.645591 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.674179 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.674156566 podStartE2EDuration="6.674156566s" podCreationTimestamp="2025-11-24 13:34:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:34:50.491567283 +0000 UTC m=+1232.131106593" watchObservedRunningTime="2025-11-24 13:34:51.674156566 +0000 UTC m=+1233.313695876" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.753094 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data-custom\") pod \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.753483 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssztq\" (UniqueName: \"kubernetes.io/projected/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-kube-api-access-ssztq\") pod \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.753524 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-logs\") pod \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.753629 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-combined-ca-bundle\") pod \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.753735 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data\") pod \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\" (UID: \"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f\") " Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.754870 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-logs" (OuterVolumeSpecName: "logs") pod "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" (UID: "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.758198 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" (UID: "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.811374 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-kube-api-access-ssztq" (OuterVolumeSpecName: "kube-api-access-ssztq") pod "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" (UID: "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f"). InnerVolumeSpecName "kube-api-access-ssztq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.857920 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" (UID: "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.858997 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.859039 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssztq\" (UniqueName: \"kubernetes.io/projected/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-kube-api-access-ssztq\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.859051 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.859061 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.885940 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data" (OuterVolumeSpecName: "config-data") pod "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" (UID: "0e6a0fbe-1914-4302-8a5d-13713a0dcc0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:51 crc kubenswrapper[4824]: I1124 13:34:51.961035 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:52 crc kubenswrapper[4824]: I1124 13:34:52.502282 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerStarted","Data":"1fb8983def1041e8782c8f361a286804573cb995da1b73740ad8fc8f182ad2b0"} Nov 24 13:34:52 crc kubenswrapper[4824]: I1124 13:34:52.506431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" event={"ID":"0e6a0fbe-1914-4302-8a5d-13713a0dcc0f","Type":"ContainerDied","Data":"a14081e678931eea8fdf708c37db8daf0451f641a9c5c4a375958b624565b352"} Nov 24 13:34:52 crc kubenswrapper[4824]: I1124 13:34:52.506545 4824 scope.go:117] "RemoveContainer" containerID="5cfe786cd566b4a539062963d92056c0917f82c46897138adcb18370f294cfc2" Nov 24 13:34:52 crc kubenswrapper[4824]: I1124 13:34:52.506776 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c8bc7b4c4-w6hnk" Nov 24 13:34:52 crc kubenswrapper[4824]: I1124 13:34:52.545422 4824 scope.go:117] "RemoveContainer" containerID="af7b85cdb856695836531bde8c6183612584f64259962d5764ae5c7e8e2af01d" Nov 24 13:34:52 crc kubenswrapper[4824]: I1124 13:34:52.555401 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c8bc7b4c4-w6hnk"] Nov 24 13:34:52 crc kubenswrapper[4824]: I1124 13:34:52.562627 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6c8bc7b4c4-w6hnk"] Nov 24 13:34:53 crc kubenswrapper[4824]: I1124 13:34:53.019645 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" path="/var/lib/kubelet/pods/0e6a0fbe-1914-4302-8a5d-13713a0dcc0f/volumes" Nov 24 13:34:53 crc kubenswrapper[4824]: I1124 13:34:53.515018 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerStarted","Data":"86df6a9256069eecef4adb29b7ec732843e733ae78a77a583cc152da7688bd7d"} Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.342106 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.408283 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-ovndb-tls-certs\") pod \"a5c8cb45-5223-4339-a957-80ec09aad094\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.408530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-httpd-config\") pod \"a5c8cb45-5223-4339-a957-80ec09aad094\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.408612 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-combined-ca-bundle\") pod \"a5c8cb45-5223-4339-a957-80ec09aad094\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.408746 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g4gw\" (UniqueName: \"kubernetes.io/projected/a5c8cb45-5223-4339-a957-80ec09aad094-kube-api-access-7g4gw\") pod \"a5c8cb45-5223-4339-a957-80ec09aad094\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.408907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-config\") pod \"a5c8cb45-5223-4339-a957-80ec09aad094\" (UID: \"a5c8cb45-5223-4339-a957-80ec09aad094\") " Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.424654 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a5c8cb45-5223-4339-a957-80ec09aad094" (UID: "a5c8cb45-5223-4339-a957-80ec09aad094"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.424943 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5c8cb45-5223-4339-a957-80ec09aad094-kube-api-access-7g4gw" (OuterVolumeSpecName: "kube-api-access-7g4gw") pod "a5c8cb45-5223-4339-a957-80ec09aad094" (UID: "a5c8cb45-5223-4339-a957-80ec09aad094"). InnerVolumeSpecName "kube-api-access-7g4gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.476292 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-config" (OuterVolumeSpecName: "config") pod "a5c8cb45-5223-4339-a957-80ec09aad094" (UID: "a5c8cb45-5223-4339-a957-80ec09aad094"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.479118 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5c8cb45-5223-4339-a957-80ec09aad094" (UID: "a5c8cb45-5223-4339-a957-80ec09aad094"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.496638 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a5c8cb45-5223-4339-a957-80ec09aad094" (UID: "a5c8cb45-5223-4339-a957-80ec09aad094"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.512950 4824 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.513424 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.513512 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.513585 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g4gw\" (UniqueName: \"kubernetes.io/projected/a5c8cb45-5223-4339-a957-80ec09aad094-kube-api-access-7g4gw\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.513641 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5c8cb45-5223-4339-a957-80ec09aad094-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.529356 4824 generic.go:334] "Generic (PLEG): container finished" podID="a5c8cb45-5223-4339-a957-80ec09aad094" containerID="c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea" exitCode=0 Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.529529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d494558c4-pqqd8" event={"ID":"a5c8cb45-5223-4339-a957-80ec09aad094","Type":"ContainerDied","Data":"c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea"} Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.530603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d494558c4-pqqd8" event={"ID":"a5c8cb45-5223-4339-a957-80ec09aad094","Type":"ContainerDied","Data":"bf28b5a6d78804dec27ada4b698f71a36861860153318a7604d2ff8fa3f02def"} Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.530706 4824 scope.go:117] "RemoveContainer" containerID="f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.529616 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d494558c4-pqqd8" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.536361 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerStarted","Data":"6656c3f9a219f361e06019a5d88b05a85b8534245bd9bcd873ade42d75c4548b"} Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.538030 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.552205 4824 scope.go:117] "RemoveContainer" containerID="c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.580732 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.85296835 podStartE2EDuration="5.580708231s" podCreationTimestamp="2025-11-24 13:34:49 +0000 UTC" firstStartedPulling="2025-11-24 13:34:50.385581294 +0000 UTC m=+1232.025120604" lastFinishedPulling="2025-11-24 13:34:54.113321165 +0000 UTC m=+1235.752860485" observedRunningTime="2025-11-24 13:34:54.560236574 +0000 UTC m=+1236.199775884" watchObservedRunningTime="2025-11-24 13:34:54.580708231 +0000 UTC m=+1236.220247561" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.585910 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d494558c4-pqqd8"] Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.589799 4824 scope.go:117] "RemoveContainer" containerID="f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56" Nov 24 13:34:54 crc kubenswrapper[4824]: E1124 13:34:54.590377 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56\": container with ID starting with f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56 not found: ID does not exist" containerID="f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.590503 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56"} err="failed to get container status \"f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56\": rpc error: code = NotFound desc = could not find container \"f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56\": container with ID starting with f17d07cfe4f1421fc8d276da18bf6017a9d30fa2f20c7ac3b56c08101cee0f56 not found: ID does not exist" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.590575 4824 scope.go:117] "RemoveContainer" containerID="c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea" Nov 24 13:34:54 crc kubenswrapper[4824]: E1124 13:34:54.591137 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea\": container with ID starting with c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea not found: ID does not exist" containerID="c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.591242 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea"} err="failed to get container status \"c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea\": rpc error: code = NotFound desc = could not find container \"c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea\": container with ID starting with c6d4504dad46332d59004e988c954f44be96cdb4608906a9d065c7fc725746ea not found: ID does not exist" Nov 24 13:34:54 crc kubenswrapper[4824]: I1124 13:34:54.597709 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7d494558c4-pqqd8"] Nov 24 13:34:55 crc kubenswrapper[4824]: I1124 13:34:55.045888 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5c8cb45-5223-4339-a957-80ec09aad094" path="/var/lib/kubelet/pods/a5c8cb45-5223-4339-a957-80ec09aad094/volumes" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.099935 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.130882 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.164010 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.225495 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-lcgn7"] Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.225725 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" podUID="6e8c3d20-1c36-4bbe-9176-3bfb989702c3" containerName="dnsmasq-dns" containerID="cri-o://7e41b82340ab499c13c31db849ba190fc51479bd1c5f1dd1b5676f412b31425d" gracePeriod=10 Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.557428 4824 generic.go:334] "Generic (PLEG): container finished" podID="6e8c3d20-1c36-4bbe-9176-3bfb989702c3" containerID="7e41b82340ab499c13c31db849ba190fc51479bd1c5f1dd1b5676f412b31425d" exitCode=0 Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.558479 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" event={"ID":"6e8c3d20-1c36-4bbe-9176-3bfb989702c3","Type":"ContainerDied","Data":"7e41b82340ab499c13c31db849ba190fc51479bd1c5f1dd1b5676f412b31425d"} Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.558612 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerName="cinder-scheduler" containerID="cri-o://47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e" gracePeriod=30 Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.559001 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerName="probe" containerID="cri-o://62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870" gracePeriod=30 Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.720277 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.863406 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-config\") pod \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.863719 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-swift-storage-0\") pod \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.864523 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-sb\") pod \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.864624 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2z9p\" (UniqueName: \"kubernetes.io/projected/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-kube-api-access-t2z9p\") pod \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.864751 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-nb\") pod \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.864858 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-svc\") pod \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.869010 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-kube-api-access-t2z9p" (OuterVolumeSpecName: "kube-api-access-t2z9p") pod "6e8c3d20-1c36-4bbe-9176-3bfb989702c3" (UID: "6e8c3d20-1c36-4bbe-9176-3bfb989702c3"). InnerVolumeSpecName "kube-api-access-t2z9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.924974 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6e8c3d20-1c36-4bbe-9176-3bfb989702c3" (UID: "6e8c3d20-1c36-4bbe-9176-3bfb989702c3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.941446 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6e8c3d20-1c36-4bbe-9176-3bfb989702c3" (UID: "6e8c3d20-1c36-4bbe-9176-3bfb989702c3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.950214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6e8c3d20-1c36-4bbe-9176-3bfb989702c3" (UID: "6e8c3d20-1c36-4bbe-9176-3bfb989702c3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:56 crc kubenswrapper[4824]: E1124 13:34:56.966166 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-nb podName:6e8c3d20-1c36-4bbe-9176-3bfb989702c3 nodeName:}" failed. No retries permitted until 2025-11-24 13:34:57.466138776 +0000 UTC m=+1239.105678086 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-nb" (UniqueName: "kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-nb") pod "6e8c3d20-1c36-4bbe-9176-3bfb989702c3" (UID: "6e8c3d20-1c36-4bbe-9176-3bfb989702c3") : error deleting /var/lib/kubelet/pods/6e8c3d20-1c36-4bbe-9176-3bfb989702c3/volume-subpaths: remove /var/lib/kubelet/pods/6e8c3d20-1c36-4bbe-9176-3bfb989702c3/volume-subpaths: no such file or directory Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.966320 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-config" (OuterVolumeSpecName: "config") pod "6e8c3d20-1c36-4bbe-9176-3bfb989702c3" (UID: "6e8c3d20-1c36-4bbe-9176-3bfb989702c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.966695 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-config\") pod \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " Nov 24 13:34:56 crc kubenswrapper[4824]: W1124 13:34:56.966853 4824 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/6e8c3d20-1c36-4bbe-9176-3bfb989702c3/volumes/kubernetes.io~configmap/config Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.966875 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-config" (OuterVolumeSpecName: "config") pod "6e8c3d20-1c36-4bbe-9176-3bfb989702c3" (UID: "6e8c3d20-1c36-4bbe-9176-3bfb989702c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.968890 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.969009 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.969111 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.969219 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:56 crc kubenswrapper[4824]: I1124 13:34:56.969488 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2z9p\" (UniqueName: \"kubernetes.io/projected/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-kube-api-access-t2z9p\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.477085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-nb\") pod \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\" (UID: \"6e8c3d20-1c36-4bbe-9176-3bfb989702c3\") " Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.477454 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6e8c3d20-1c36-4bbe-9176-3bfb989702c3" (UID: "6e8c3d20-1c36-4bbe-9176-3bfb989702c3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.477844 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e8c3d20-1c36-4bbe-9176-3bfb989702c3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.573970 4824 generic.go:334] "Generic (PLEG): container finished" podID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerID="62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870" exitCode=0 Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.574044 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9","Type":"ContainerDied","Data":"62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870"} Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.601336 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" event={"ID":"6e8c3d20-1c36-4bbe-9176-3bfb989702c3","Type":"ContainerDied","Data":"fe1abd3bba1d5aa7974ad7372061220c105d76c55549ba3e8c92376c1024818b"} Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.601625 4824 scope.go:117] "RemoveContainer" containerID="7e41b82340ab499c13c31db849ba190fc51479bd1c5f1dd1b5676f412b31425d" Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.601776 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-lcgn7" Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.628076 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-lcgn7"] Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.633592 4824 scope.go:117] "RemoveContainer" containerID="7cf598350f47422e877c652160c24677cce39a786288ffb70dc49ca4b5475592" Nov 24 13:34:57 crc kubenswrapper[4824]: I1124 13:34:57.638510 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-lcgn7"] Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.520072 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.597448 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.611540 4824 generic.go:334] "Generic (PLEG): container finished" podID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerID="47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e" exitCode=0 Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.611602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9","Type":"ContainerDied","Data":"47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e"} Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.611636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9","Type":"ContainerDied","Data":"fc656c368945e6bdebbea039132045aa8c5879a94bf444d3f16c10bfbfca877f"} Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.611653 4824 scope.go:117] "RemoveContainer" containerID="62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.612262 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.621654 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zbz6\" (UniqueName: \"kubernetes.io/projected/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-kube-api-access-2zbz6\") pod \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.621901 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data\") pod \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.622071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-combined-ca-bundle\") pod \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.622216 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-etc-machine-id\") pod \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.622287 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-scripts\") pod \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.622421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data-custom\") pod \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\" (UID: \"4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9\") " Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.622616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" (UID: "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.622898 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.671681 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-kube-api-access-2zbz6" (OuterVolumeSpecName: "kube-api-access-2zbz6") pod "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" (UID: "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9"). InnerVolumeSpecName "kube-api-access-2zbz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.675791 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-scripts" (OuterVolumeSpecName: "scripts") pod "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" (UID: "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.681731 4824 scope.go:117] "RemoveContainer" containerID="47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.684122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" (UID: "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.727691 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.727718 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zbz6\" (UniqueName: \"kubernetes.io/projected/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-kube-api-access-2zbz6\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.727728 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.756162 4824 scope.go:117] "RemoveContainer" containerID="62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870" Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.759859 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870\": container with ID starting with 62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870 not found: ID does not exist" containerID="62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.760013 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870"} err="failed to get container status \"62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870\": rpc error: code = NotFound desc = could not find container \"62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870\": container with ID starting with 62fa660b4a0eb03eb6bed9a61bef97607f9016862f52ea8790431e1abfdf6870 not found: ID does not exist" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.760104 4824 scope.go:117] "RemoveContainer" containerID="47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e" Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.761462 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e\": container with ID starting with 47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e not found: ID does not exist" containerID="47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.761699 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e"} err="failed to get container status \"47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e\": rpc error: code = NotFound desc = could not find container \"47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e\": container with ID starting with 47f01278886f749b2553ea57a9897ba860ee6974782497330d65e4c3299e7d4e not found: ID does not exist" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.763892 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" (UID: "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.816912 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data" (OuterVolumeSpecName: "config-data") pod "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" (UID: "4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.829704 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.829734 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.941454 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.954027 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970329 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.970651 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api-log" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970671 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api-log" Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.970684 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970691 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api" Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.970704 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerName="cinder-scheduler" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970710 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerName="cinder-scheduler" Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.970724 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c8cb45-5223-4339-a957-80ec09aad094" containerName="neutron-api" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970729 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c8cb45-5223-4339-a957-80ec09aad094" containerName="neutron-api" Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.970743 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c8cb45-5223-4339-a957-80ec09aad094" containerName="neutron-httpd" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970751 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c8cb45-5223-4339-a957-80ec09aad094" containerName="neutron-httpd" Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.970766 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e8c3d20-1c36-4bbe-9176-3bfb989702c3" containerName="dnsmasq-dns" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970771 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e8c3d20-1c36-4bbe-9176-3bfb989702c3" containerName="dnsmasq-dns" Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.970778 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e8c3d20-1c36-4bbe-9176-3bfb989702c3" containerName="init" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970783 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e8c3d20-1c36-4bbe-9176-3bfb989702c3" containerName="init" Nov 24 13:34:58 crc kubenswrapper[4824]: E1124 13:34:58.970792 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerName="probe" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970798 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerName="probe" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970960 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerName="probe" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970970 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e8c3d20-1c36-4bbe-9176-3bfb989702c3" containerName="dnsmasq-dns" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970984 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c8cb45-5223-4339-a957-80ec09aad094" containerName="neutron-api" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.970999 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api-log" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.971012 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c8cb45-5223-4339-a957-80ec09aad094" containerName="neutron-httpd" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.971030 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" containerName="cinder-scheduler" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.971040 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6a0fbe-1914-4302-8a5d-13713a0dcc0f" containerName="barbican-api" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.972000 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.974735 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 13:34:58 crc kubenswrapper[4824]: I1124 13:34:58.982640 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.021167 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9" path="/var/lib/kubelet/pods/4b34abdc-f5c1-427d-bb72-a6efbd1bbcb9/volumes" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.021753 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e8c3d20-1c36-4bbe-9176-3bfb989702c3" path="/var/lib/kubelet/pods/6e8c3d20-1c36-4bbe-9176-3bfb989702c3/volumes" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.134348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-config-data\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.134405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp4vz\" (UniqueName: \"kubernetes.io/projected/51aacb45-4a5f-42d3-b905-a051a14856f4-kube-api-access-lp4vz\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.134447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.134495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.134511 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-scripts\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.134534 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51aacb45-4a5f-42d3-b905-a051a14856f4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.236376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp4vz\" (UniqueName: \"kubernetes.io/projected/51aacb45-4a5f-42d3-b905-a051a14856f4-kube-api-access-lp4vz\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.236444 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.236500 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.236518 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-scripts\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.236556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51aacb45-4a5f-42d3-b905-a051a14856f4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.236613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-config-data\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.236842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51aacb45-4a5f-42d3-b905-a051a14856f4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.240888 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-config-data\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.241006 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.241364 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.245314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51aacb45-4a5f-42d3-b905-a051a14856f4-scripts\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.253062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp4vz\" (UniqueName: \"kubernetes.io/projected/51aacb45-4a5f-42d3-b905-a051a14856f4-kube-api-access-lp4vz\") pod \"cinder-scheduler-0\" (UID: \"51aacb45-4a5f-42d3-b905-a051a14856f4\") " pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.290254 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 13:34:59 crc kubenswrapper[4824]: I1124 13:34:59.742970 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 13:35:00 crc kubenswrapper[4824]: I1124 13:35:00.511648 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:35:00 crc kubenswrapper[4824]: I1124 13:35:00.596994 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-bdb7b6678-pnfdp" Nov 24 13:35:00 crc kubenswrapper[4824]: I1124 13:35:00.660908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51aacb45-4a5f-42d3-b905-a051a14856f4","Type":"ContainerStarted","Data":"4dfc6e14cb1cfe802ebee374ef10300911a5ad7d6f691193f844ced85100a9b3"} Nov 24 13:35:00 crc kubenswrapper[4824]: I1124 13:35:00.661196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51aacb45-4a5f-42d3-b905-a051a14856f4","Type":"ContainerStarted","Data":"676a40a0eba6093b9b70664cb43f10a95d4d5e084cb73a2f9ae3171dbf760f3d"} Nov 24 13:35:00 crc kubenswrapper[4824]: I1124 13:35:00.848909 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-57d8799bcb-x76nj" Nov 24 13:35:01 crc kubenswrapper[4824]: I1124 13:35:01.668241 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51aacb45-4a5f-42d3-b905-a051a14856f4","Type":"ContainerStarted","Data":"96a3e818504cb0af24ad28c966bc4c8abfc8175ccb4526ccff4e69fb8438cb0d"} Nov 24 13:35:01 crc kubenswrapper[4824]: I1124 13:35:01.691518 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.691501267 podStartE2EDuration="3.691501267s" podCreationTimestamp="2025-11-24 13:34:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:01.686468818 +0000 UTC m=+1243.326008118" watchObservedRunningTime="2025-11-24 13:35:01.691501267 +0000 UTC m=+1243.331040597" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.559905 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.561854 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.564223 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-jjm7f" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.564638 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.564937 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.580917 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.743914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-openstack-config-secret\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.743965 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-openstack-config\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.744126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.744162 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qml5w\" (UniqueName: \"kubernetes.io/projected/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-kube-api-access-qml5w\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.846045 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.846094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qml5w\" (UniqueName: \"kubernetes.io/projected/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-kube-api-access-qml5w\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.846197 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-openstack-config-secret\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.846218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-openstack-config\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.846995 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-openstack-config\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.851856 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.858260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-openstack-config-secret\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.874489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qml5w\" (UniqueName: \"kubernetes.io/projected/7d1c1940-0b8b-4b09-ae2a-76168c5a873b-kube-api-access-qml5w\") pod \"openstackclient\" (UID: \"7d1c1940-0b8b-4b09-ae2a-76168c5a873b\") " pod="openstack/openstackclient" Nov 24 13:35:02 crc kubenswrapper[4824]: I1124 13:35:02.881536 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 13:35:03 crc kubenswrapper[4824]: I1124 13:35:03.483686 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 13:35:03 crc kubenswrapper[4824]: I1124 13:35:03.702050 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"7d1c1940-0b8b-4b09-ae2a-76168c5a873b","Type":"ContainerStarted","Data":"8e420604747be7dbdb54948e8874a8637fe96da9eef98d6aeb53b82c8d9f51d1"} Nov 24 13:35:04 crc kubenswrapper[4824]: I1124 13:35:04.290896 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.462887 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.463601 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="ceilometer-central-agent" containerID="cri-o://3de161309266940d53a6d44c8aecccb0b7d77ddf2238234eedd6c4003797aa69" gracePeriod=30 Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.463768 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="sg-core" containerID="cri-o://86df6a9256069eecef4adb29b7ec732843e733ae78a77a583cc152da7688bd7d" gracePeriod=30 Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.463827 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="ceilometer-notification-agent" containerID="cri-o://1fb8983def1041e8782c8f361a286804573cb995da1b73740ad8fc8f182ad2b0" gracePeriod=30 Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.463772 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="proxy-httpd" containerID="cri-o://6656c3f9a219f361e06019a5d88b05a85b8534245bd9bcd873ade42d75c4548b" gracePeriod=30 Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.481612 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.746263 4824 generic.go:334] "Generic (PLEG): container finished" podID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerID="6656c3f9a219f361e06019a5d88b05a85b8534245bd9bcd873ade42d75c4548b" exitCode=0 Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.746291 4824 generic.go:334] "Generic (PLEG): container finished" podID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerID="86df6a9256069eecef4adb29b7ec732843e733ae78a77a583cc152da7688bd7d" exitCode=2 Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.746310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerDied","Data":"6656c3f9a219f361e06019a5d88b05a85b8534245bd9bcd873ade42d75c4548b"} Nov 24 13:35:06 crc kubenswrapper[4824]: I1124 13:35:06.746335 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerDied","Data":"86df6a9256069eecef4adb29b7ec732843e733ae78a77a583cc152da7688bd7d"} Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.081173 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-8cc59fd67-2j94h"] Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.084302 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.093922 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.094156 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.095312 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.101133 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-8cc59fd67-2j94h"] Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.227027 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-internal-tls-certs\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.227083 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-run-httpd\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.227132 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-config-data\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.227153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-log-httpd\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.227188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-public-tls-certs\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.227214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l474g\" (UniqueName: \"kubernetes.io/projected/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-kube-api-access-l474g\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.227251 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-etc-swift\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.227286 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-combined-ca-bundle\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.239535 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.242494 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-log" containerID="cri-o://460c47f4638325b913b7972be871cd2beee6d07988efb6ce1b04ef745b1130d0" gracePeriod=30 Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.243006 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-httpd" containerID="cri-o://d1a92abb4c37ed9efbc60fa33188f3ff303d3d8678908e453a7329176942773b" gracePeriod=30 Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.328480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l474g\" (UniqueName: \"kubernetes.io/projected/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-kube-api-access-l474g\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.328545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-etc-swift\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.328589 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-combined-ca-bundle\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.328631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-internal-tls-certs\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.328667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-run-httpd\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.328720 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-config-data\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.328741 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-log-httpd\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.328776 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-public-tls-certs\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.330270 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-log-httpd\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.330559 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-run-httpd\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.337390 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-config-data\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.338631 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-combined-ca-bundle\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.341857 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-public-tls-certs\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.341906 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-internal-tls-certs\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.358011 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-etc-swift\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.366601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l474g\" (UniqueName: \"kubernetes.io/projected/57b08eb1-c1bf-4b56-880f-4d8adeaa32fc-kube-api-access-l474g\") pod \"swift-proxy-8cc59fd67-2j94h\" (UID: \"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc\") " pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.410551 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.773771 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerID="460c47f4638325b913b7972be871cd2beee6d07988efb6ce1b04ef745b1130d0" exitCode=143 Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.774050 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b60d128-15e0-4646-ad33-c562d28aedbc","Type":"ContainerDied","Data":"460c47f4638325b913b7972be871cd2beee6d07988efb6ce1b04ef745b1130d0"} Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.778771 4824 generic.go:334] "Generic (PLEG): container finished" podID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerID="3de161309266940d53a6d44c8aecccb0b7d77ddf2238234eedd6c4003797aa69" exitCode=0 Nov 24 13:35:07 crc kubenswrapper[4824]: I1124 13:35:07.778835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerDied","Data":"3de161309266940d53a6d44c8aecccb0b7d77ddf2238234eedd6c4003797aa69"} Nov 24 13:35:08 crc kubenswrapper[4824]: I1124 13:35:08.137436 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-8cc59fd67-2j94h"] Nov 24 13:35:08 crc kubenswrapper[4824]: I1124 13:35:08.789288 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8cc59fd67-2j94h" event={"ID":"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc","Type":"ContainerStarted","Data":"da44fd019f34a0b66ecd07997d7b5021356bead10f7153030ab16b2a824443f3"} Nov 24 13:35:08 crc kubenswrapper[4824]: I1124 13:35:08.789521 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8cc59fd67-2j94h" event={"ID":"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc","Type":"ContainerStarted","Data":"67a46f252cf4fb77299c0c45c3a7a7ca30b5b87d0c938ef5ed99114b3f4eba94"} Nov 24 13:35:08 crc kubenswrapper[4824]: I1124 13:35:08.789534 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8cc59fd67-2j94h" event={"ID":"57b08eb1-c1bf-4b56-880f-4d8adeaa32fc","Type":"ContainerStarted","Data":"7a3b9f3779dcb8df3a664d8182d5f1d481f92fa516aa862406d238ba54a8e1c8"} Nov 24 13:35:08 crc kubenswrapper[4824]: I1124 13:35:08.789658 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:08 crc kubenswrapper[4824]: I1124 13:35:08.810956 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-8cc59fd67-2j94h" podStartSLOduration=1.8109389660000001 podStartE2EDuration="1.810938966s" podCreationTimestamp="2025-11-24 13:35:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:08.805823525 +0000 UTC m=+1250.445362835" watchObservedRunningTime="2025-11-24 13:35:08.810938966 +0000 UTC m=+1250.450478276" Nov 24 13:35:09 crc kubenswrapper[4824]: I1124 13:35:09.543042 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 13:35:09 crc kubenswrapper[4824]: I1124 13:35:09.803271 4824 generic.go:334] "Generic (PLEG): container finished" podID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerID="1fb8983def1041e8782c8f361a286804573cb995da1b73740ad8fc8f182ad2b0" exitCode=0 Nov 24 13:35:09 crc kubenswrapper[4824]: I1124 13:35:09.803356 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerDied","Data":"1fb8983def1041e8782c8f361a286804573cb995da1b73740ad8fc8f182ad2b0"} Nov 24 13:35:09 crc kubenswrapper[4824]: I1124 13:35:09.803550 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:10 crc kubenswrapper[4824]: I1124 13:35:10.787893 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:35:10 crc kubenswrapper[4824]: I1124 13:35:10.788202 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:35:10 crc kubenswrapper[4824]: I1124 13:35:10.814897 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerID="d1a92abb4c37ed9efbc60fa33188f3ff303d3d8678908e453a7329176942773b" exitCode=0 Nov 24 13:35:10 crc kubenswrapper[4824]: I1124 13:35:10.814948 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b60d128-15e0-4646-ad33-c562d28aedbc","Type":"ContainerDied","Data":"d1a92abb4c37ed9efbc60fa33188f3ff303d3d8678908e453a7329176942773b"} Nov 24 13:35:14 crc kubenswrapper[4824]: I1124 13:35:14.381002 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:35:14 crc kubenswrapper[4824]: I1124 13:35:14.381694 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-log" containerID="cri-o://3abafe77dceb2eda0505bc39ac04ec7eda6553623108aa3cdedfb8ec7c03da03" gracePeriod=30 Nov 24 13:35:14 crc kubenswrapper[4824]: I1124 13:35:14.381844 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-httpd" containerID="cri-o://93a38e62aa44c86cd016c05ce04fc6b0cf9264567f9ff24a5340836da177bed2" gracePeriod=30 Nov 24 13:35:14 crc kubenswrapper[4824]: I1124 13:35:14.628709 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.151:9292/healthcheck\": dial tcp 10.217.0.151:9292: connect: connection refused" Nov 24 13:35:14 crc kubenswrapper[4824]: I1124 13:35:14.629027 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.151:9292/healthcheck\": dial tcp 10.217.0.151:9292: connect: connection refused" Nov 24 13:35:14 crc kubenswrapper[4824]: I1124 13:35:14.851501 4824 generic.go:334] "Generic (PLEG): container finished" podID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerID="3abafe77dceb2eda0505bc39ac04ec7eda6553623108aa3cdedfb8ec7c03da03" exitCode=143 Nov 24 13:35:14 crc kubenswrapper[4824]: I1124 13:35:14.851569 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"36a320b3-eeb2-4400-8d0e-27360ab167ee","Type":"ContainerDied","Data":"3abafe77dceb2eda0505bc39ac04ec7eda6553623108aa3cdedfb8ec7c03da03"} Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.821687 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.892238 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.895042 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-log-httpd\") pod \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.895087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-scripts\") pod \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.895136 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-config-data\") pod \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.895199 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-combined-ca-bundle\") pod \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.895240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-run-httpd\") pod \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.895301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh7p4\" (UniqueName: \"kubernetes.io/projected/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-kube-api-access-kh7p4\") pod \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.895373 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-sg-core-conf-yaml\") pod \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\" (UID: \"2738c9b8-9700-46d5-aa31-fb1e760ffc2e\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.896039 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2738c9b8-9700-46d5-aa31-fb1e760ffc2e" (UID: "2738c9b8-9700-46d5-aa31-fb1e760ffc2e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.896099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2738c9b8-9700-46d5-aa31-fb1e760ffc2e" (UID: "2738c9b8-9700-46d5-aa31-fb1e760ffc2e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.907586 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-scripts" (OuterVolumeSpecName: "scripts") pod "2738c9b8-9700-46d5-aa31-fb1e760ffc2e" (UID: "2738c9b8-9700-46d5-aa31-fb1e760ffc2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.910966 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-kube-api-access-kh7p4" (OuterVolumeSpecName: "kube-api-access-kh7p4") pod "2738c9b8-9700-46d5-aa31-fb1e760ffc2e" (UID: "2738c9b8-9700-46d5-aa31-fb1e760ffc2e"). InnerVolumeSpecName "kube-api-access-kh7p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.926867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2738c9b8-9700-46d5-aa31-fb1e760ffc2e","Type":"ContainerDied","Data":"85843d22616999c9e1c9eaa8ac48981ed7834b1d4ae28e612b71973855591fed"} Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.927027 4824 scope.go:117] "RemoveContainer" containerID="6656c3f9a219f361e06019a5d88b05a85b8534245bd9bcd873ade42d75c4548b" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.927215 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.947138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"7d1c1940-0b8b-4b09-ae2a-76168c5a873b","Type":"ContainerStarted","Data":"270b41617ad3595aed29e8d61b46899251e7397a0f8f01ab7dc7b2b7dda0251e"} Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.950757 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2738c9b8-9700-46d5-aa31-fb1e760ffc2e" (UID: "2738c9b8-9700-46d5-aa31-fb1e760ffc2e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.968967 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b60d128-15e0-4646-ad33-c562d28aedbc","Type":"ContainerDied","Data":"b40991aa3544468f25ea0be0078985ffc0980bc424f1240478a3d08510a2bf6a"} Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.969049 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.977306 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.115667421 podStartE2EDuration="13.977286273s" podCreationTimestamp="2025-11-24 13:35:02 +0000 UTC" firstStartedPulling="2025-11-24 13:35:03.51329131 +0000 UTC m=+1245.152830620" lastFinishedPulling="2025-11-24 13:35:15.374910162 +0000 UTC m=+1257.014449472" observedRunningTime="2025-11-24 13:35:15.961398694 +0000 UTC m=+1257.600938004" watchObservedRunningTime="2025-11-24 13:35:15.977286273 +0000 UTC m=+1257.616825573" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.996865 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-scripts\") pod \"6b60d128-15e0-4646-ad33-c562d28aedbc\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.996944 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-config-data\") pod \"6b60d128-15e0-4646-ad33-c562d28aedbc\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.996986 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-logs\") pod \"6b60d128-15e0-4646-ad33-c562d28aedbc\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-public-tls-certs\") pod \"6b60d128-15e0-4646-ad33-c562d28aedbc\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997127 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-combined-ca-bundle\") pod \"6b60d128-15e0-4646-ad33-c562d28aedbc\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997179 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-httpd-run\") pod \"6b60d128-15e0-4646-ad33-c562d28aedbc\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997208 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fh4q\" (UniqueName: \"kubernetes.io/projected/6b60d128-15e0-4646-ad33-c562d28aedbc-kube-api-access-4fh4q\") pod \"6b60d128-15e0-4646-ad33-c562d28aedbc\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997230 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"6b60d128-15e0-4646-ad33-c562d28aedbc\" (UID: \"6b60d128-15e0-4646-ad33-c562d28aedbc\") " Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997762 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh7p4\" (UniqueName: \"kubernetes.io/projected/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-kube-api-access-kh7p4\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997788 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997834 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997849 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.997859 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.998717 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-logs" (OuterVolumeSpecName: "logs") pod "6b60d128-15e0-4646-ad33-c562d28aedbc" (UID: "6b60d128-15e0-4646-ad33-c562d28aedbc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.998951 4824 scope.go:117] "RemoveContainer" containerID="86df6a9256069eecef4adb29b7ec732843e733ae78a77a583cc152da7688bd7d" Nov 24 13:35:15 crc kubenswrapper[4824]: I1124 13:35:15.999282 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6b60d128-15e0-4646-ad33-c562d28aedbc" (UID: "6b60d128-15e0-4646-ad33-c562d28aedbc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.008012 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b60d128-15e0-4646-ad33-c562d28aedbc-kube-api-access-4fh4q" (OuterVolumeSpecName: "kube-api-access-4fh4q") pod "6b60d128-15e0-4646-ad33-c562d28aedbc" (UID: "6b60d128-15e0-4646-ad33-c562d28aedbc"). InnerVolumeSpecName "kube-api-access-4fh4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.023323 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "6b60d128-15e0-4646-ad33-c562d28aedbc" (UID: "6b60d128-15e0-4646-ad33-c562d28aedbc"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.028880 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-scripts" (OuterVolumeSpecName: "scripts") pod "6b60d128-15e0-4646-ad33-c562d28aedbc" (UID: "6b60d128-15e0-4646-ad33-c562d28aedbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.053256 4824 scope.go:117] "RemoveContainer" containerID="1fb8983def1041e8782c8f361a286804573cb995da1b73740ad8fc8f182ad2b0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.064589 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b60d128-15e0-4646-ad33-c562d28aedbc" (UID: "6b60d128-15e0-4646-ad33-c562d28aedbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.078394 4824 scope.go:117] "RemoveContainer" containerID="3de161309266940d53a6d44c8aecccb0b7d77ddf2238234eedd6c4003797aa69" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.099187 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.099220 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.099232 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.099242 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b60d128-15e0-4646-ad33-c562d28aedbc-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.099257 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fh4q\" (UniqueName: \"kubernetes.io/projected/6b60d128-15e0-4646-ad33-c562d28aedbc-kube-api-access-4fh4q\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.099279 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.113489 4824 scope.go:117] "RemoveContainer" containerID="d1a92abb4c37ed9efbc60fa33188f3ff303d3d8678908e453a7329176942773b" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.145205 4824 scope.go:117] "RemoveContainer" containerID="460c47f4638325b913b7972be871cd2beee6d07988efb6ce1b04ef745b1130d0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.146149 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-config-data" (OuterVolumeSpecName: "config-data") pod "6b60d128-15e0-4646-ad33-c562d28aedbc" (UID: "6b60d128-15e0-4646-ad33-c562d28aedbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.147058 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.161052 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6b60d128-15e0-4646-ad33-c562d28aedbc" (UID: "6b60d128-15e0-4646-ad33-c562d28aedbc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.161139 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2738c9b8-9700-46d5-aa31-fb1e760ffc2e" (UID: "2738c9b8-9700-46d5-aa31-fb1e760ffc2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.165178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-config-data" (OuterVolumeSpecName: "config-data") pod "2738c9b8-9700-46d5-aa31-fb1e760ffc2e" (UID: "2738c9b8-9700-46d5-aa31-fb1e760ffc2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.201225 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.201263 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b60d128-15e0-4646-ad33-c562d28aedbc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.201277 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.201288 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.201300 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2738c9b8-9700-46d5-aa31-fb1e760ffc2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.354125 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.365785 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.373918 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.384998 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392061 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:16 crc kubenswrapper[4824]: E1124 13:35:16.392435 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-log" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392455 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-log" Nov 24 13:35:16 crc kubenswrapper[4824]: E1124 13:35:16.392482 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-httpd" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392490 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-httpd" Nov 24 13:35:16 crc kubenswrapper[4824]: E1124 13:35:16.392510 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="ceilometer-central-agent" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392517 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="ceilometer-central-agent" Nov 24 13:35:16 crc kubenswrapper[4824]: E1124 13:35:16.392545 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="sg-core" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392553 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="sg-core" Nov 24 13:35:16 crc kubenswrapper[4824]: E1124 13:35:16.392564 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="ceilometer-notification-agent" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392570 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="ceilometer-notification-agent" Nov 24 13:35:16 crc kubenswrapper[4824]: E1124 13:35:16.392584 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="proxy-httpd" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392589 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="proxy-httpd" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392855 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="ceilometer-central-agent" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392881 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="proxy-httpd" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.392990 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="ceilometer-notification-agent" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.393007 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-log" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.393021 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" containerName="glance-httpd" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.393031 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" containerName="sg-core" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.395048 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.410417 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.412964 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.413875 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.414552 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.420310 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.420512 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.432853 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.440288 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kqq5\" (UniqueName: \"kubernetes.io/projected/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-kube-api-access-4kqq5\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518255 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxshq\" (UniqueName: \"kubernetes.io/projected/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-kube-api-access-mxshq\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518278 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-scripts\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-config-data\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518497 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-log-httpd\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518643 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518672 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-logs\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518710 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518819 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-run-httpd\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518882 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.518925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-config-data\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.519058 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-scripts\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.620978 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-log-httpd\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-logs\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621093 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621121 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621139 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-run-httpd\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621172 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621195 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-config-data\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621237 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-scripts\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621283 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kqq5\" (UniqueName: \"kubernetes.io/projected/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-kube-api-access-4kqq5\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621298 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxshq\" (UniqueName: \"kubernetes.io/projected/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-kube-api-access-mxshq\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621316 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-scripts\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621361 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-config-data\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.621898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-logs\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.622322 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.623280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.624422 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-log-httpd\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.624743 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-run-httpd\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.630003 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.632219 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-scripts\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.633082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.633085 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-config-data\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.636684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.637785 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.638109 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-config-data\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.640480 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-scripts\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.643127 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kqq5\" (UniqueName: \"kubernetes.io/projected/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-kube-api-access-4kqq5\") pod \"ceilometer-0\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.651111 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxshq\" (UniqueName: \"kubernetes.io/projected/f7321bd0-b6b4-45dc-a4c6-c7a26b601b17-kube-api-access-mxshq\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.670003 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17\") " pod="openstack/glance-default-external-api-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.724880 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:16 crc kubenswrapper[4824]: I1124 13:35:16.735819 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 13:35:17 crc kubenswrapper[4824]: I1124 13:35:17.032621 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2738c9b8-9700-46d5-aa31-fb1e760ffc2e" path="/var/lib/kubelet/pods/2738c9b8-9700-46d5-aa31-fb1e760ffc2e/volumes" Nov 24 13:35:17 crc kubenswrapper[4824]: I1124 13:35:17.033902 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b60d128-15e0-4646-ad33-c562d28aedbc" path="/var/lib/kubelet/pods/6b60d128-15e0-4646-ad33-c562d28aedbc/volumes" Nov 24 13:35:17 crc kubenswrapper[4824]: I1124 13:35:17.180902 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:17 crc kubenswrapper[4824]: I1124 13:35:17.291456 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:17 crc kubenswrapper[4824]: W1124 13:35:17.310859 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb11f5b64_d3ec_4c7b_abad_3ab9af4fc48c.slice/crio-cbbb31c8104af9d0def7592ac916a6d0b691ef64db39306c12899ab9095c1632 WatchSource:0}: Error finding container cbbb31c8104af9d0def7592ac916a6d0b691ef64db39306c12899ab9095c1632: Status 404 returned error can't find the container with id cbbb31c8104af9d0def7592ac916a6d0b691ef64db39306c12899ab9095c1632 Nov 24 13:35:17 crc kubenswrapper[4824]: I1124 13:35:17.350447 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 13:35:17 crc kubenswrapper[4824]: W1124 13:35:17.358842 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7321bd0_b6b4_45dc_a4c6_c7a26b601b17.slice/crio-c22a7e7234dbd9fd2769a0d7f4c2a131a32b6647a8ed0e1776825d49da0983ac WatchSource:0}: Error finding container c22a7e7234dbd9fd2769a0d7f4c2a131a32b6647a8ed0e1776825d49da0983ac: Status 404 returned error can't find the container with id c22a7e7234dbd9fd2769a0d7f4c2a131a32b6647a8ed0e1776825d49da0983ac Nov 24 13:35:17 crc kubenswrapper[4824]: I1124 13:35:17.420762 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:17 crc kubenswrapper[4824]: I1124 13:35:17.422562 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-8cc59fd67-2j94h" Nov 24 13:35:17 crc kubenswrapper[4824]: I1124 13:35:17.580462 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": read tcp 10.217.0.2:36506->10.217.0.150:9292: read: connection reset by peer" Nov 24 13:35:17 crc kubenswrapper[4824]: I1124 13:35:17.580826 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": read tcp 10.217.0.2:36490->10.217.0.150:9292: read: connection reset by peer" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.026686 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerStarted","Data":"cbbb31c8104af9d0def7592ac916a6d0b691ef64db39306c12899ab9095c1632"} Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.029933 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17","Type":"ContainerStarted","Data":"c22a7e7234dbd9fd2769a0d7f4c2a131a32b6647a8ed0e1776825d49da0983ac"} Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.035440 4824 generic.go:334] "Generic (PLEG): container finished" podID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerID="93a38e62aa44c86cd016c05ce04fc6b0cf9264567f9ff24a5340836da177bed2" exitCode=0 Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.035677 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"36a320b3-eeb2-4400-8d0e-27360ab167ee","Type":"ContainerDied","Data":"93a38e62aa44c86cd016c05ce04fc6b0cf9264567f9ff24a5340836da177bed2"} Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.199302 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.313434 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-internal-tls-certs\") pod \"36a320b3-eeb2-4400-8d0e-27360ab167ee\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.313503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-combined-ca-bundle\") pod \"36a320b3-eeb2-4400-8d0e-27360ab167ee\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.313562 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-logs\") pod \"36a320b3-eeb2-4400-8d0e-27360ab167ee\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.313592 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-scripts\") pod \"36a320b3-eeb2-4400-8d0e-27360ab167ee\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.313630 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-config-data\") pod \"36a320b3-eeb2-4400-8d0e-27360ab167ee\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.313677 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4qmp\" (UniqueName: \"kubernetes.io/projected/36a320b3-eeb2-4400-8d0e-27360ab167ee-kube-api-access-q4qmp\") pod \"36a320b3-eeb2-4400-8d0e-27360ab167ee\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.313713 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-httpd-run\") pod \"36a320b3-eeb2-4400-8d0e-27360ab167ee\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.313766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"36a320b3-eeb2-4400-8d0e-27360ab167ee\" (UID: \"36a320b3-eeb2-4400-8d0e-27360ab167ee\") " Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.322636 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-logs" (OuterVolumeSpecName: "logs") pod "36a320b3-eeb2-4400-8d0e-27360ab167ee" (UID: "36a320b3-eeb2-4400-8d0e-27360ab167ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.325178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "36a320b3-eeb2-4400-8d0e-27360ab167ee" (UID: "36a320b3-eeb2-4400-8d0e-27360ab167ee"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.328524 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "36a320b3-eeb2-4400-8d0e-27360ab167ee" (UID: "36a320b3-eeb2-4400-8d0e-27360ab167ee"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.334983 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36a320b3-eeb2-4400-8d0e-27360ab167ee-kube-api-access-q4qmp" (OuterVolumeSpecName: "kube-api-access-q4qmp") pod "36a320b3-eeb2-4400-8d0e-27360ab167ee" (UID: "36a320b3-eeb2-4400-8d0e-27360ab167ee"). InnerVolumeSpecName "kube-api-access-q4qmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.339402 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-scripts" (OuterVolumeSpecName: "scripts") pod "36a320b3-eeb2-4400-8d0e-27360ab167ee" (UID: "36a320b3-eeb2-4400-8d0e-27360ab167ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.349081 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36a320b3-eeb2-4400-8d0e-27360ab167ee" (UID: "36a320b3-eeb2-4400-8d0e-27360ab167ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.415353 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4qmp\" (UniqueName: \"kubernetes.io/projected/36a320b3-eeb2-4400-8d0e-27360ab167ee-kube-api-access-q4qmp\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.415386 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.415461 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.415472 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.415482 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36a320b3-eeb2-4400-8d0e-27360ab167ee-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.415490 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.472502 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.490980 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-config-data" (OuterVolumeSpecName: "config-data") pod "36a320b3-eeb2-4400-8d0e-27360ab167ee" (UID: "36a320b3-eeb2-4400-8d0e-27360ab167ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.516532 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.516713 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.573111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "36a320b3-eeb2-4400-8d0e-27360ab167ee" (UID: "36a320b3-eeb2-4400-8d0e-27360ab167ee"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:18 crc kubenswrapper[4824]: I1124 13:35:18.620163 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36a320b3-eeb2-4400-8d0e-27360ab167ee-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.060949 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"36a320b3-eeb2-4400-8d0e-27360ab167ee","Type":"ContainerDied","Data":"029e8aada7d7da37038c43aa7bea540f12a8ba8032df61ee7851bc6f481d1f25"} Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.061240 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.065776 4824 scope.go:117] "RemoveContainer" containerID="93a38e62aa44c86cd016c05ce04fc6b0cf9264567f9ff24a5340836da177bed2" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.074567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerStarted","Data":"c28806afd6e12758d4c4d4b34c38780219ec2417c2429f5a218b1f0ef84eac33"} Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.074835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerStarted","Data":"a5d2267d08330494487cbc0479e6988d0f056ac32cd5d5e8654839193e72a8fe"} Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.079090 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17","Type":"ContainerStarted","Data":"9c9574721cba4fb994c00d7f4a5b09b5ec630b0d184a40780a43013fba0d2524"} Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.079195 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7321bd0-b6b4-45dc-a4c6-c7a26b601b17","Type":"ContainerStarted","Data":"4d516d64cabb79b24d9cc72a26eed1cb4f96b885da2b35b8ad60017293c2fdd7"} Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.117786 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.127515 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.129871 4824 scope.go:117] "RemoveContainer" containerID="3abafe77dceb2eda0505bc39ac04ec7eda6553623108aa3cdedfb8ec7c03da03" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.151744 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:35:19 crc kubenswrapper[4824]: E1124 13:35:19.152133 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-log" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.152150 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-log" Nov 24 13:35:19 crc kubenswrapper[4824]: E1124 13:35:19.152167 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-httpd" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.152174 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-httpd" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.152365 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-log" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.152400 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" containerName="glance-httpd" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.174153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.183025 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.184104 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.194721 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.198769 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.198750448 podStartE2EDuration="3.198750448s" podCreationTimestamp="2025-11-24 13:35:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:19.133702463 +0000 UTC m=+1260.773241783" watchObservedRunningTime="2025-11-24 13:35:19.198750448 +0000 UTC m=+1260.838289758" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.231860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.231914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.231957 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.231998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.232023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.232073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnm5f\" (UniqueName: \"kubernetes.io/projected/dda48846-9717-4dd4-b4af-afe6887923bb-kube-api-access-lnm5f\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.232144 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dda48846-9717-4dd4-b4af-afe6887923bb-logs\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.232172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dda48846-9717-4dd4-b4af-afe6887923bb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.334677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dda48846-9717-4dd4-b4af-afe6887923bb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.334788 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.334880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.334916 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.335123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dda48846-9717-4dd4-b4af-afe6887923bb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.335609 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.335638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.335666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnm5f\" (UniqueName: \"kubernetes.io/projected/dda48846-9717-4dd4-b4af-afe6887923bb-kube-api-access-lnm5f\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.335742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dda48846-9717-4dd4-b4af-afe6887923bb-logs\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.337379 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.338120 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dda48846-9717-4dd4-b4af-afe6887923bb-logs\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.343939 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.345066 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.354644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnm5f\" (UniqueName: \"kubernetes.io/projected/dda48846-9717-4dd4-b4af-afe6887923bb-kube-api-access-lnm5f\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.356439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.356505 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dda48846-9717-4dd4-b4af-afe6887923bb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.385348 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"dda48846-9717-4dd4-b4af-afe6887923bb\") " pod="openstack/glance-default-internal-api-0" Nov 24 13:35:19 crc kubenswrapper[4824]: I1124 13:35:19.499699 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:20 crc kubenswrapper[4824]: I1124 13:35:20.083041 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 13:35:20 crc kubenswrapper[4824]: I1124 13:35:20.088230 4824 generic.go:334] "Generic (PLEG): container finished" podID="5828a51f-d146-44ea-b8ea-423e017939d8" containerID="f0eb7c3f4f829b0e7b8112bc730b12912e7eb168f9eb8c9618b6934afbce96bd" exitCode=137 Nov 24 13:35:20 crc kubenswrapper[4824]: I1124 13:35:20.088288 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-959bb6dd8-8pcrj" event={"ID":"5828a51f-d146-44ea-b8ea-423e017939d8","Type":"ContainerDied","Data":"f0eb7c3f4f829b0e7b8112bc730b12912e7eb168f9eb8c9618b6934afbce96bd"} Nov 24 13:35:20 crc kubenswrapper[4824]: I1124 13:35:20.088314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-959bb6dd8-8pcrj" event={"ID":"5828a51f-d146-44ea-b8ea-423e017939d8","Type":"ContainerStarted","Data":"10153c59514b1fb9d9bd5d60a607d78b4fef73886b3292eecbba8bc41c9c727e"} Nov 24 13:35:20 crc kubenswrapper[4824]: I1124 13:35:20.092609 4824 generic.go:334] "Generic (PLEG): container finished" podID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerID="6ce765b74f266b883a095df6fbfd4992669f9774df16a85346b8570f4fc1ede0" exitCode=137 Nov 24 13:35:20 crc kubenswrapper[4824]: I1124 13:35:20.092672 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c96f4646d-wsbgv" event={"ID":"bc2d2fae-b656-464e-b1a8-89b4c5160792","Type":"ContainerDied","Data":"6ce765b74f266b883a095df6fbfd4992669f9774df16a85346b8570f4fc1ede0"} Nov 24 13:35:20 crc kubenswrapper[4824]: I1124 13:35:20.092713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c96f4646d-wsbgv" event={"ID":"bc2d2fae-b656-464e-b1a8-89b4c5160792","Type":"ContainerStarted","Data":"1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f"} Nov 24 13:35:20 crc kubenswrapper[4824]: I1124 13:35:20.094556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerStarted","Data":"70f02ac7a73545f8f53279a1dbf9d88bbf6f48226e632a0cf632a1086a8bd920"} Nov 24 13:35:20 crc kubenswrapper[4824]: W1124 13:35:20.097943 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddda48846_9717_4dd4_b4af_afe6887923bb.slice/crio-5f1f34ded1fe28c65d68c52ccc03382f9b782f3020471041118127c63168aada WatchSource:0}: Error finding container 5f1f34ded1fe28c65d68c52ccc03382f9b782f3020471041118127c63168aada: Status 404 returned error can't find the container with id 5f1f34ded1fe28c65d68c52ccc03382f9b782f3020471041118127c63168aada Nov 24 13:35:20 crc kubenswrapper[4824]: I1124 13:35:20.972015 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.044932 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36a320b3-eeb2-4400-8d0e-27360ab167ee" path="/var/lib/kubelet/pods/36a320b3-eeb2-4400-8d0e-27360ab167ee/volumes" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.082731 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50ecbaf8-dbbf-4d2c-8394-b0319768e871-etc-machine-id\") pod \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.082871 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data\") pod \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.082917 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-scripts\") pod \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.082983 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data-custom\") pod \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.083027 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-combined-ca-bundle\") pod \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.083066 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50ecbaf8-dbbf-4d2c-8394-b0319768e871-logs\") pod \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.083081 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd4x9\" (UniqueName: \"kubernetes.io/projected/50ecbaf8-dbbf-4d2c-8394-b0319768e871-kube-api-access-vd4x9\") pod \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\" (UID: \"50ecbaf8-dbbf-4d2c-8394-b0319768e871\") " Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.084068 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50ecbaf8-dbbf-4d2c-8394-b0319768e871-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "50ecbaf8-dbbf-4d2c-8394-b0319768e871" (UID: "50ecbaf8-dbbf-4d2c-8394-b0319768e871"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.086947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50ecbaf8-dbbf-4d2c-8394-b0319768e871-logs" (OuterVolumeSpecName: "logs") pod "50ecbaf8-dbbf-4d2c-8394-b0319768e871" (UID: "50ecbaf8-dbbf-4d2c-8394-b0319768e871"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.105159 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50ecbaf8-dbbf-4d2c-8394-b0319768e871-kube-api-access-vd4x9" (OuterVolumeSpecName: "kube-api-access-vd4x9") pod "50ecbaf8-dbbf-4d2c-8394-b0319768e871" (UID: "50ecbaf8-dbbf-4d2c-8394-b0319768e871"). InnerVolumeSpecName "kube-api-access-vd4x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.109934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-scripts" (OuterVolumeSpecName: "scripts") pod "50ecbaf8-dbbf-4d2c-8394-b0319768e871" (UID: "50ecbaf8-dbbf-4d2c-8394-b0319768e871"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.123503 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "50ecbaf8-dbbf-4d2c-8394-b0319768e871" (UID: "50ecbaf8-dbbf-4d2c-8394-b0319768e871"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.151212 4824 generic.go:334] "Generic (PLEG): container finished" podID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerID="c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff" exitCode=137 Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.151276 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50ecbaf8-dbbf-4d2c-8394-b0319768e871","Type":"ContainerDied","Data":"c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff"} Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.151302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50ecbaf8-dbbf-4d2c-8394-b0319768e871","Type":"ContainerDied","Data":"6ab0f3ebf4e9f0109c7f7e9b5c7261ec4dabc1f1051c9cc9732cd404124cacc2"} Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.151318 4824 scope.go:117] "RemoveContainer" containerID="c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.151414 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.168132 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dda48846-9717-4dd4-b4af-afe6887923bb","Type":"ContainerStarted","Data":"fcd6a087d32a35483ecc2c9729389a03cc439ea9f106f166c0c05744ee173601"} Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.168182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dda48846-9717-4dd4-b4af-afe6887923bb","Type":"ContainerStarted","Data":"5f1f34ded1fe28c65d68c52ccc03382f9b782f3020471041118127c63168aada"} Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.187037 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50ecbaf8-dbbf-4d2c-8394-b0319768e871" (UID: "50ecbaf8-dbbf-4d2c-8394-b0319768e871"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.188299 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.188325 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.188335 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.188344 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50ecbaf8-dbbf-4d2c-8394-b0319768e871-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.188352 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd4x9\" (UniqueName: \"kubernetes.io/projected/50ecbaf8-dbbf-4d2c-8394-b0319768e871-kube-api-access-vd4x9\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.188360 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50ecbaf8-dbbf-4d2c-8394-b0319768e871-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.190909 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerStarted","Data":"5188c15d99318df99f49785675db3ab51338bc05d856fd6a71280708191d9e9f"} Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.191063 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="ceilometer-central-agent" containerID="cri-o://a5d2267d08330494487cbc0479e6988d0f056ac32cd5d5e8654839193e72a8fe" gracePeriod=30 Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.191340 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.192775 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="ceilometer-notification-agent" containerID="cri-o://c28806afd6e12758d4c4d4b34c38780219ec2417c2429f5a218b1f0ef84eac33" gracePeriod=30 Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.192832 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="sg-core" containerID="cri-o://70f02ac7a73545f8f53279a1dbf9d88bbf6f48226e632a0cf632a1086a8bd920" gracePeriod=30 Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.192959 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="proxy-httpd" containerID="cri-o://5188c15d99318df99f49785675db3ab51338bc05d856fd6a71280708191d9e9f" gracePeriod=30 Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.212090 4824 scope.go:117] "RemoveContainer" containerID="17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.227091 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data" (OuterVolumeSpecName: "config-data") pod "50ecbaf8-dbbf-4d2c-8394-b0319768e871" (UID: "50ecbaf8-dbbf-4d2c-8394-b0319768e871"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.248432 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.853928799 podStartE2EDuration="5.248414768s" podCreationTimestamp="2025-11-24 13:35:16 +0000 UTC" firstStartedPulling="2025-11-24 13:35:17.313918993 +0000 UTC m=+1258.953458303" lastFinishedPulling="2025-11-24 13:35:20.708404962 +0000 UTC m=+1262.347944272" observedRunningTime="2025-11-24 13:35:21.233131795 +0000 UTC m=+1262.872671115" watchObservedRunningTime="2025-11-24 13:35:21.248414768 +0000 UTC m=+1262.887954078" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.291226 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ecbaf8-dbbf-4d2c-8394-b0319768e871-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.309021 4824 scope.go:117] "RemoveContainer" containerID="c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff" Nov 24 13:35:21 crc kubenswrapper[4824]: E1124 13:35:21.330945 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff\": container with ID starting with c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff not found: ID does not exist" containerID="c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.330989 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff"} err="failed to get container status \"c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff\": rpc error: code = NotFound desc = could not find container \"c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff\": container with ID starting with c3cf1d675f61eb5431ceb0006776cdc108ed068bdcbaf5ef4b8ee68d5d86ddff not found: ID does not exist" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.331013 4824 scope.go:117] "RemoveContainer" containerID="17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274" Nov 24 13:35:21 crc kubenswrapper[4824]: E1124 13:35:21.338360 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274\": container with ID starting with 17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274 not found: ID does not exist" containerID="17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.338498 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274"} err="failed to get container status \"17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274\": rpc error: code = NotFound desc = could not find container \"17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274\": container with ID starting with 17b3815f3955c2a7054cecfe69ffa9ea745e33ff853c58038328427ea2975274 not found: ID does not exist" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.502369 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.510923 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.531583 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:35:21 crc kubenswrapper[4824]: E1124 13:35:21.532047 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerName="cinder-api-log" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.532075 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerName="cinder-api-log" Nov 24 13:35:21 crc kubenswrapper[4824]: E1124 13:35:21.532105 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerName="cinder-api" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.532114 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerName="cinder-api" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.532337 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerName="cinder-api" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.532368 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" containerName="cinder-api-log" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.533435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.542444 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.542621 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.542718 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.562210 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.597235 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.597291 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.597333 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4xvd\" (UniqueName: \"kubernetes.io/projected/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-kube-api-access-d4xvd\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.597380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-config-data\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.597407 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-scripts\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.597440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.597463 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-logs\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.597487 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.597545 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-config-data-custom\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.700630 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.700676 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.700723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4xvd\" (UniqueName: \"kubernetes.io/projected/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-kube-api-access-d4xvd\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.700752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-config-data\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.700779 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-scripts\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.700797 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.700827 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-logs\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.700846 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.700883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-config-data-custom\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.705268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-logs\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.708539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.708795 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.710038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.710209 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-config-data\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.710403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-config-data-custom\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.719210 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-scripts\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.725064 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.725919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4xvd\" (UniqueName: \"kubernetes.io/projected/4e42a7c2-0db7-4c5d-b731-b543c875ebb0-kube-api-access-d4xvd\") pod \"cinder-api-0\" (UID: \"4e42a7c2-0db7-4c5d-b731-b543c875ebb0\") " pod="openstack/cinder-api-0" Nov 24 13:35:21 crc kubenswrapper[4824]: I1124 13:35:21.868037 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 13:35:22 crc kubenswrapper[4824]: I1124 13:35:22.208968 4824 generic.go:334] "Generic (PLEG): container finished" podID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerID="5188c15d99318df99f49785675db3ab51338bc05d856fd6a71280708191d9e9f" exitCode=0 Nov 24 13:35:22 crc kubenswrapper[4824]: I1124 13:35:22.209235 4824 generic.go:334] "Generic (PLEG): container finished" podID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerID="70f02ac7a73545f8f53279a1dbf9d88bbf6f48226e632a0cf632a1086a8bd920" exitCode=2 Nov 24 13:35:22 crc kubenswrapper[4824]: I1124 13:35:22.209244 4824 generic.go:334] "Generic (PLEG): container finished" podID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerID="c28806afd6e12758d4c4d4b34c38780219ec2417c2429f5a218b1f0ef84eac33" exitCode=0 Nov 24 13:35:22 crc kubenswrapper[4824]: I1124 13:35:22.209286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerDied","Data":"5188c15d99318df99f49785675db3ab51338bc05d856fd6a71280708191d9e9f"} Nov 24 13:35:22 crc kubenswrapper[4824]: I1124 13:35:22.209310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerDied","Data":"70f02ac7a73545f8f53279a1dbf9d88bbf6f48226e632a0cf632a1086a8bd920"} Nov 24 13:35:22 crc kubenswrapper[4824]: I1124 13:35:22.209319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerDied","Data":"c28806afd6e12758d4c4d4b34c38780219ec2417c2429f5a218b1f0ef84eac33"} Nov 24 13:35:22 crc kubenswrapper[4824]: I1124 13:35:22.211157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dda48846-9717-4dd4-b4af-afe6887923bb","Type":"ContainerStarted","Data":"fc483ad0226d256cfc14c773c4dc15207288bd5f47cd0f0f6885014e2d959f32"} Nov 24 13:35:22 crc kubenswrapper[4824]: I1124 13:35:22.230066 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.230050705 podStartE2EDuration="3.230050705s" podCreationTimestamp="2025-11-24 13:35:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:22.226465143 +0000 UTC m=+1263.866004453" watchObservedRunningTime="2025-11-24 13:35:22.230050705 +0000 UTC m=+1263.869590015" Nov 24 13:35:22 crc kubenswrapper[4824]: I1124 13:35:22.394299 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 13:35:23 crc kubenswrapper[4824]: I1124 13:35:23.022124 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50ecbaf8-dbbf-4d2c-8394-b0319768e871" path="/var/lib/kubelet/pods/50ecbaf8-dbbf-4d2c-8394-b0319768e871/volumes" Nov 24 13:35:23 crc kubenswrapper[4824]: I1124 13:35:23.225756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4e42a7c2-0db7-4c5d-b731-b543c875ebb0","Type":"ContainerStarted","Data":"3cf30bcb79152297b5739af3c8ed6fa36a6afbc1083b3d34f53659fbdc3650ed"} Nov 24 13:35:23 crc kubenswrapper[4824]: I1124 13:35:23.226174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4e42a7c2-0db7-4c5d-b731-b543c875ebb0","Type":"ContainerStarted","Data":"7ec5d0087d304acf08117537c9549fe75b77e83c944c4a01973272beacb59682"} Nov 24 13:35:24 crc kubenswrapper[4824]: I1124 13:35:24.234432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4e42a7c2-0db7-4c5d-b731-b543c875ebb0","Type":"ContainerStarted","Data":"5ff390a05492540492583337a15756dabe8e92cf3fd78c78ce6ade0c9ddea63d"} Nov 24 13:35:24 crc kubenswrapper[4824]: I1124 13:35:24.234848 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 13:35:24 crc kubenswrapper[4824]: I1124 13:35:24.252396 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.252378712 podStartE2EDuration="3.252378712s" podCreationTimestamp="2025-11-24 13:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:24.249045646 +0000 UTC m=+1265.888584956" watchObservedRunningTime="2025-11-24 13:35:24.252378712 +0000 UTC m=+1265.891918022" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.260503 4824 generic.go:334] "Generic (PLEG): container finished" podID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerID="a5d2267d08330494487cbc0479e6988d0f056ac32cd5d5e8654839193e72a8fe" exitCode=0 Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.260688 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerDied","Data":"a5d2267d08330494487cbc0479e6988d0f056ac32cd5d5e8654839193e72a8fe"} Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.485840 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.611753 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-scripts\") pod \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.611950 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kqq5\" (UniqueName: \"kubernetes.io/projected/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-kube-api-access-4kqq5\") pod \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.612097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-run-httpd\") pod \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.612132 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-combined-ca-bundle\") pod \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.612186 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-config-data\") pod \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.612334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-sg-core-conf-yaml\") pod \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.612425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-log-httpd\") pod \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\" (UID: \"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c\") " Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.612335 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" (UID: "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.613062 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" (UID: "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.613077 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.617953 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-kube-api-access-4kqq5" (OuterVolumeSpecName: "kube-api-access-4kqq5") pod "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" (UID: "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c"). InnerVolumeSpecName "kube-api-access-4kqq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.648040 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-scripts" (OuterVolumeSpecName: "scripts") pod "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" (UID: "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.662661 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" (UID: "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.697925 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" (UID: "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.714870 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.714900 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.714909 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.714918 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kqq5\" (UniqueName: \"kubernetes.io/projected/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-kube-api-access-4kqq5\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.714928 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.736755 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.736852 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.741384 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-config-data" (OuterVolumeSpecName: "config-data") pod "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" (UID: "b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.777325 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.777856 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 13:35:26 crc kubenswrapper[4824]: I1124 13:35:26.817255 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.308147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c","Type":"ContainerDied","Data":"cbbb31c8104af9d0def7592ac916a6d0b691ef64db39306c12899ab9095c1632"} Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.308217 4824 scope.go:117] "RemoveContainer" containerID="5188c15d99318df99f49785675db3ab51338bc05d856fd6a71280708191d9e9f" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.308426 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.309116 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.309147 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.337457 4824 scope.go:117] "RemoveContainer" containerID="70f02ac7a73545f8f53279a1dbf9d88bbf6f48226e632a0cf632a1086a8bd920" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.342238 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.349460 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.359978 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:27 crc kubenswrapper[4824]: E1124 13:35:27.360355 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="ceilometer-central-agent" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.360374 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="ceilometer-central-agent" Nov 24 13:35:27 crc kubenswrapper[4824]: E1124 13:35:27.360405 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="ceilometer-notification-agent" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.360413 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="ceilometer-notification-agent" Nov 24 13:35:27 crc kubenswrapper[4824]: E1124 13:35:27.360436 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="sg-core" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.360443 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="sg-core" Nov 24 13:35:27 crc kubenswrapper[4824]: E1124 13:35:27.360457 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="proxy-httpd" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.360466 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="proxy-httpd" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.360640 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="sg-core" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.360659 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="ceilometer-central-agent" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.360678 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="ceilometer-notification-agent" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.360690 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" containerName="proxy-httpd" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.368619 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.375904 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.376179 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.384593 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.428211 4824 scope.go:117] "RemoveContainer" containerID="c28806afd6e12758d4c4d4b34c38780219ec2417c2429f5a218b1f0ef84eac33" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.456498 4824 scope.go:117] "RemoveContainer" containerID="a5d2267d08330494487cbc0479e6988d0f056ac32cd5d5e8654839193e72a8fe" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.538664 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-run-httpd\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.539025 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-config-data\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.539218 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.539463 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.539621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-scripts\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.539740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hnbh\" (UniqueName: \"kubernetes.io/projected/ff67241a-4c14-4ebf-913a-69e8a5beb5af-kube-api-access-4hnbh\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.539909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-log-httpd\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.642138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.642445 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.642494 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-scripts\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.642522 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hnbh\" (UniqueName: \"kubernetes.io/projected/ff67241a-4c14-4ebf-913a-69e8a5beb5af-kube-api-access-4hnbh\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.642555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-log-httpd\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.642597 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-run-httpd\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.642623 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-config-data\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.643312 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-log-httpd\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.643375 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-run-httpd\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.647231 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-scripts\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.647370 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.653651 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-config-data\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.658513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.659577 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hnbh\" (UniqueName: \"kubernetes.io/projected/ff67241a-4c14-4ebf-913a-69e8a5beb5af-kube-api-access-4hnbh\") pod \"ceilometer-0\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " pod="openstack/ceilometer-0" Nov 24 13:35:27 crc kubenswrapper[4824]: I1124 13:35:27.687066 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:28 crc kubenswrapper[4824]: I1124 13:35:28.207749 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:28 crc kubenswrapper[4824]: I1124 13:35:28.316715 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerStarted","Data":"0ad7cf68605235b6d92551147f520938342b5ce3877070bbca07054be7f7b563"} Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.067743 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c" path="/var/lib/kubelet/pods/b11f5b64-d3ec-4c7b-abad-3ab9af4fc48c/volumes" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.328130 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.328418 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.329196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerStarted","Data":"0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa"} Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.466472 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.466872 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.468028 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.501427 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.501466 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.543440 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.562973 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.671282 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:35:29 crc kubenswrapper[4824]: I1124 13:35:29.671336 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:35:30 crc kubenswrapper[4824]: I1124 13:35:30.237580 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 13:35:30 crc kubenswrapper[4824]: I1124 13:35:30.276124 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 13:35:30 crc kubenswrapper[4824]: I1124 13:35:30.340106 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerStarted","Data":"7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d"} Nov 24 13:35:30 crc kubenswrapper[4824]: I1124 13:35:30.341227 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:30 crc kubenswrapper[4824]: I1124 13:35:30.341304 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:31 crc kubenswrapper[4824]: I1124 13:35:31.353312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerStarted","Data":"ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f"} Nov 24 13:35:32 crc kubenswrapper[4824]: I1124 13:35:32.362981 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:35:32 crc kubenswrapper[4824]: I1124 13:35:32.363213 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:35:33 crc kubenswrapper[4824]: I1124 13:35:33.026936 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:33 crc kubenswrapper[4824]: I1124 13:35:33.158793 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 13:35:33 crc kubenswrapper[4824]: I1124 13:35:33.373493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerStarted","Data":"53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04"} Nov 24 13:35:33 crc kubenswrapper[4824]: I1124 13:35:33.373551 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:35:33 crc kubenswrapper[4824]: I1124 13:35:33.895415 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.736272134 podStartE2EDuration="6.895396914s" podCreationTimestamp="2025-11-24 13:35:27 +0000 UTC" firstStartedPulling="2025-11-24 13:35:28.214484258 +0000 UTC m=+1269.854023568" lastFinishedPulling="2025-11-24 13:35:32.373609038 +0000 UTC m=+1274.013148348" observedRunningTime="2025-11-24 13:35:33.402129392 +0000 UTC m=+1275.041668702" watchObservedRunningTime="2025-11-24 13:35:33.895396914 +0000 UTC m=+1275.534936224" Nov 24 13:35:33 crc kubenswrapper[4824]: I1124 13:35:33.903274 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:35 crc kubenswrapper[4824]: I1124 13:35:35.375261 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 13:35:35 crc kubenswrapper[4824]: I1124 13:35:35.385539 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="ceilometer-central-agent" containerID="cri-o://0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa" gracePeriod=30 Nov 24 13:35:35 crc kubenswrapper[4824]: I1124 13:35:35.385551 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="sg-core" containerID="cri-o://ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f" gracePeriod=30 Nov 24 13:35:35 crc kubenswrapper[4824]: I1124 13:35:35.385593 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="ceilometer-notification-agent" containerID="cri-o://7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d" gracePeriod=30 Nov 24 13:35:35 crc kubenswrapper[4824]: I1124 13:35:35.385557 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="proxy-httpd" containerID="cri-o://53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04" gracePeriod=30 Nov 24 13:35:36 crc kubenswrapper[4824]: I1124 13:35:36.396418 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerID="53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04" exitCode=0 Nov 24 13:35:36 crc kubenswrapper[4824]: I1124 13:35:36.396766 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerID="ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f" exitCode=2 Nov 24 13:35:36 crc kubenswrapper[4824]: I1124 13:35:36.396777 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerID="7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d" exitCode=0 Nov 24 13:35:36 crc kubenswrapper[4824]: I1124 13:35:36.396523 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerDied","Data":"53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04"} Nov 24 13:35:36 crc kubenswrapper[4824]: I1124 13:35:36.396830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerDied","Data":"ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f"} Nov 24 13:35:36 crc kubenswrapper[4824]: I1124 13:35:36.396846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerDied","Data":"7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d"} Nov 24 13:35:37 crc kubenswrapper[4824]: I1124 13:35:37.853736 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-mnhn5"] Nov 24 13:35:37 crc kubenswrapper[4824]: I1124 13:35:37.856211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:37 crc kubenswrapper[4824]: I1124 13:35:37.885476 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mnhn5"] Nov 24 13:35:37 crc kubenswrapper[4824]: I1124 13:35:37.949924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc09e1d-e018-4086-b17e-f7a1ccf02951-operator-scripts\") pod \"nova-api-db-create-mnhn5\" (UID: \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\") " pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:37 crc kubenswrapper[4824]: I1124 13:35:37.950040 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jx98\" (UniqueName: \"kubernetes.io/projected/8fc09e1d-e018-4086-b17e-f7a1ccf02951-kube-api-access-5jx98\") pod \"nova-api-db-create-mnhn5\" (UID: \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\") " pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.012329 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-5h6sl"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.013520 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.033468 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-5h6sl"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.049916 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f0d5-account-create-46mmw"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.051314 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.051537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jx98\" (UniqueName: \"kubernetes.io/projected/8fc09e1d-e018-4086-b17e-f7a1ccf02951-kube-api-access-5jx98\") pod \"nova-api-db-create-mnhn5\" (UID: \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\") " pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.051618 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc09e1d-e018-4086-b17e-f7a1ccf02951-operator-scripts\") pod \"nova-api-db-create-mnhn5\" (UID: \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\") " pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.052746 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc09e1d-e018-4086-b17e-f7a1ccf02951-operator-scripts\") pod \"nova-api-db-create-mnhn5\" (UID: \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\") " pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.057694 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.075535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jx98\" (UniqueName: \"kubernetes.io/projected/8fc09e1d-e018-4086-b17e-f7a1ccf02951-kube-api-access-5jx98\") pod \"nova-api-db-create-mnhn5\" (UID: \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\") " pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.100446 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f0d5-account-create-46mmw"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.154513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f396378e-6a0d-4853-a622-3d5ca30d45fc-operator-scripts\") pod \"nova-cell0-db-create-5h6sl\" (UID: \"f396378e-6a0d-4853-a622-3d5ca30d45fc\") " pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.154676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570330a6-c219-4a73-8731-b85ca896c82c-operator-scripts\") pod \"nova-api-f0d5-account-create-46mmw\" (UID: \"570330a6-c219-4a73-8731-b85ca896c82c\") " pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.154719 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddc76\" (UniqueName: \"kubernetes.io/projected/570330a6-c219-4a73-8731-b85ca896c82c-kube-api-access-ddc76\") pod \"nova-api-f0d5-account-create-46mmw\" (UID: \"570330a6-c219-4a73-8731-b85ca896c82c\") " pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.154765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2lr9\" (UniqueName: \"kubernetes.io/projected/f396378e-6a0d-4853-a622-3d5ca30d45fc-kube-api-access-b2lr9\") pod \"nova-cell0-db-create-5h6sl\" (UID: \"f396378e-6a0d-4853-a622-3d5ca30d45fc\") " pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.172050 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.249398 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0a19-account-create-gxxbb"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.250726 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.254459 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.256882 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddc76\" (UniqueName: \"kubernetes.io/projected/570330a6-c219-4a73-8731-b85ca896c82c-kube-api-access-ddc76\") pod \"nova-api-f0d5-account-create-46mmw\" (UID: \"570330a6-c219-4a73-8731-b85ca896c82c\") " pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.257195 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2lr9\" (UniqueName: \"kubernetes.io/projected/f396378e-6a0d-4853-a622-3d5ca30d45fc-kube-api-access-b2lr9\") pod \"nova-cell0-db-create-5h6sl\" (UID: \"f396378e-6a0d-4853-a622-3d5ca30d45fc\") " pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.257267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f396378e-6a0d-4853-a622-3d5ca30d45fc-operator-scripts\") pod \"nova-cell0-db-create-5h6sl\" (UID: \"f396378e-6a0d-4853-a622-3d5ca30d45fc\") " pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.257357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570330a6-c219-4a73-8731-b85ca896c82c-operator-scripts\") pod \"nova-api-f0d5-account-create-46mmw\" (UID: \"570330a6-c219-4a73-8731-b85ca896c82c\") " pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.258139 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570330a6-c219-4a73-8731-b85ca896c82c-operator-scripts\") pod \"nova-api-f0d5-account-create-46mmw\" (UID: \"570330a6-c219-4a73-8731-b85ca896c82c\") " pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.258750 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f396378e-6a0d-4853-a622-3d5ca30d45fc-operator-scripts\") pod \"nova-cell0-db-create-5h6sl\" (UID: \"f396378e-6a0d-4853-a622-3d5ca30d45fc\") " pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.266916 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-8dxxk"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.268061 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.292901 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0a19-account-create-gxxbb"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.295161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2lr9\" (UniqueName: \"kubernetes.io/projected/f396378e-6a0d-4853-a622-3d5ca30d45fc-kube-api-access-b2lr9\") pod \"nova-cell0-db-create-5h6sl\" (UID: \"f396378e-6a0d-4853-a622-3d5ca30d45fc\") " pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.327139 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.346326 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-8dxxk"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.349533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddc76\" (UniqueName: \"kubernetes.io/projected/570330a6-c219-4a73-8731-b85ca896c82c-kube-api-access-ddc76\") pod \"nova-api-f0d5-account-create-46mmw\" (UID: \"570330a6-c219-4a73-8731-b85ca896c82c\") " pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.367644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2ba678-6efb-4ec5-b163-da752d2edecb-operator-scripts\") pod \"nova-cell0-0a19-account-create-gxxbb\" (UID: \"9f2ba678-6efb-4ec5-b163-da752d2edecb\") " pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.367782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbvr5\" (UniqueName: \"kubernetes.io/projected/6cbf4587-2c54-4220-a514-62b043882b5b-kube-api-access-gbvr5\") pod \"nova-cell1-db-create-8dxxk\" (UID: \"6cbf4587-2c54-4220-a514-62b043882b5b\") " pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.367837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cbf4587-2c54-4220-a514-62b043882b5b-operator-scripts\") pod \"nova-cell1-db-create-8dxxk\" (UID: \"6cbf4587-2c54-4220-a514-62b043882b5b\") " pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.367872 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhfd8\" (UniqueName: \"kubernetes.io/projected/9f2ba678-6efb-4ec5-b163-da752d2edecb-kube-api-access-zhfd8\") pod \"nova-cell0-0a19-account-create-gxxbb\" (UID: \"9f2ba678-6efb-4ec5-b163-da752d2edecb\") " pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.368271 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.474401 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbvr5\" (UniqueName: \"kubernetes.io/projected/6cbf4587-2c54-4220-a514-62b043882b5b-kube-api-access-gbvr5\") pod \"nova-cell1-db-create-8dxxk\" (UID: \"6cbf4587-2c54-4220-a514-62b043882b5b\") " pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.474454 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cbf4587-2c54-4220-a514-62b043882b5b-operator-scripts\") pod \"nova-cell1-db-create-8dxxk\" (UID: \"6cbf4587-2c54-4220-a514-62b043882b5b\") " pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.474489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhfd8\" (UniqueName: \"kubernetes.io/projected/9f2ba678-6efb-4ec5-b163-da752d2edecb-kube-api-access-zhfd8\") pod \"nova-cell0-0a19-account-create-gxxbb\" (UID: \"9f2ba678-6efb-4ec5-b163-da752d2edecb\") " pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.474534 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2ba678-6efb-4ec5-b163-da752d2edecb-operator-scripts\") pod \"nova-cell0-0a19-account-create-gxxbb\" (UID: \"9f2ba678-6efb-4ec5-b163-da752d2edecb\") " pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.475442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2ba678-6efb-4ec5-b163-da752d2edecb-operator-scripts\") pod \"nova-cell0-0a19-account-create-gxxbb\" (UID: \"9f2ba678-6efb-4ec5-b163-da752d2edecb\") " pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.476161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cbf4587-2c54-4220-a514-62b043882b5b-operator-scripts\") pod \"nova-cell1-db-create-8dxxk\" (UID: \"6cbf4587-2c54-4220-a514-62b043882b5b\") " pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.517166 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbvr5\" (UniqueName: \"kubernetes.io/projected/6cbf4587-2c54-4220-a514-62b043882b5b-kube-api-access-gbvr5\") pod \"nova-cell1-db-create-8dxxk\" (UID: \"6cbf4587-2c54-4220-a514-62b043882b5b\") " pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.525384 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhfd8\" (UniqueName: \"kubernetes.io/projected/9f2ba678-6efb-4ec5-b163-da752d2edecb-kube-api-access-zhfd8\") pod \"nova-cell0-0a19-account-create-gxxbb\" (UID: \"9f2ba678-6efb-4ec5-b163-da752d2edecb\") " pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.557765 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-e59f-account-create-t5q2j"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.561611 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.572596 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.613342 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e59f-account-create-t5q2j"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.691610 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-operator-scripts\") pod \"nova-cell1-e59f-account-create-t5q2j\" (UID: \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\") " pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.691648 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76skc\" (UniqueName: \"kubernetes.io/projected/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-kube-api-access-76skc\") pod \"nova-cell1-e59f-account-create-t5q2j\" (UID: \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\") " pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.730406 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.745653 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.793552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-operator-scripts\") pod \"nova-cell1-e59f-account-create-t5q2j\" (UID: \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\") " pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.793598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76skc\" (UniqueName: \"kubernetes.io/projected/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-kube-api-access-76skc\") pod \"nova-cell1-e59f-account-create-t5q2j\" (UID: \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\") " pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.795231 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-operator-scripts\") pod \"nova-cell1-e59f-account-create-t5q2j\" (UID: \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\") " pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.817419 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76skc\" (UniqueName: \"kubernetes.io/projected/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-kube-api-access-76skc\") pod \"nova-cell1-e59f-account-create-t5q2j\" (UID: \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\") " pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.888366 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-5h6sl"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.903471 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mnhn5"] Nov 24 13:35:38 crc kubenswrapper[4824]: I1124 13:35:38.943599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.209861 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f0d5-account-create-46mmw"] Nov 24 13:35:39 crc kubenswrapper[4824]: W1124 13:35:39.247028 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod570330a6_c219_4a73_8731_b85ca896c82c.slice/crio-4d84a830f209a9251d4495a67ab029422fb60a97c707919bc451a4c9790cf5a3 WatchSource:0}: Error finding container 4d84a830f209a9251d4495a67ab029422fb60a97c707919bc451a4c9790cf5a3: Status 404 returned error can't find the container with id 4d84a830f209a9251d4495a67ab029422fb60a97c707919bc451a4c9790cf5a3 Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.476120 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 24 13:35:39 crc kubenswrapper[4824]: W1124 13:35:39.502858 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f2ba678_6efb_4ec5_b163_da752d2edecb.slice/crio-bc6abe9a21ee62df1fb65f2e7eaa4d90982b043964559e1dce19e37ea68c9ad9 WatchSource:0}: Error finding container bc6abe9a21ee62df1fb65f2e7eaa4d90982b043964559e1dce19e37ea68c9ad9: Status 404 returned error can't find the container with id bc6abe9a21ee62df1fb65f2e7eaa4d90982b043964559e1dce19e37ea68c9ad9 Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.534723 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0a19-account-create-gxxbb"] Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.549400 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f0d5-account-create-46mmw" event={"ID":"570330a6-c219-4a73-8731-b85ca896c82c","Type":"ContainerStarted","Data":"4d84a830f209a9251d4495a67ab029422fb60a97c707919bc451a4c9790cf5a3"} Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.551355 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5h6sl" event={"ID":"f396378e-6a0d-4853-a622-3d5ca30d45fc","Type":"ContainerStarted","Data":"096465ccc50d35cad02903956880aa23ed4855202d5dbbf17089b3b383417e33"} Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.551385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5h6sl" event={"ID":"f396378e-6a0d-4853-a622-3d5ca30d45fc","Type":"ContainerStarted","Data":"6ab3f96d7e8d5df22a0f733667582627a49e0bbaec68d8f8a39f675fa53793a2"} Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.572871 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mnhn5" event={"ID":"8fc09e1d-e018-4086-b17e-f7a1ccf02951","Type":"ContainerStarted","Data":"46bb48c4039626b4768b3567f43f875aec9a2f905b9013e4919fccbcf98ffd9f"} Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.585201 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-5h6sl" podStartSLOduration=2.5851800689999997 podStartE2EDuration="2.585180069s" podCreationTimestamp="2025-11-24 13:35:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:39.577692466 +0000 UTC m=+1281.217231776" watchObservedRunningTime="2025-11-24 13:35:39.585180069 +0000 UTC m=+1281.224719379" Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.663398 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-8dxxk"] Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.681974 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-959bb6dd8-8pcrj" podUID="5828a51f-d146-44ea-b8ea-423e017939d8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 24 13:35:39 crc kubenswrapper[4824]: I1124 13:35:39.872869 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e59f-account-create-t5q2j"] Nov 24 13:35:39 crc kubenswrapper[4824]: W1124 13:35:39.879924 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e9dd708_d69f_4c23_bbe1_2f41c7f18fae.slice/crio-95c237dade31fb7fb8f50f086a9bedba7b28d79a4a45b4e81a7ee9adfc91d695 WatchSource:0}: Error finding container 95c237dade31fb7fb8f50f086a9bedba7b28d79a4a45b4e81a7ee9adfc91d695: Status 404 returned error can't find the container with id 95c237dade31fb7fb8f50f086a9bedba7b28d79a4a45b4e81a7ee9adfc91d695 Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.585647 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f0d5-account-create-46mmw" event={"ID":"570330a6-c219-4a73-8731-b85ca896c82c","Type":"ContainerStarted","Data":"f1d7def3cc42ff723e398a77610016b05a96420dd0b60e3e0395650b2a042e20"} Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.588712 4824 generic.go:334] "Generic (PLEG): container finished" podID="f396378e-6a0d-4853-a622-3d5ca30d45fc" containerID="096465ccc50d35cad02903956880aa23ed4855202d5dbbf17089b3b383417e33" exitCode=0 Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.588853 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5h6sl" event={"ID":"f396378e-6a0d-4853-a622-3d5ca30d45fc","Type":"ContainerDied","Data":"096465ccc50d35cad02903956880aa23ed4855202d5dbbf17089b3b383417e33"} Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.591348 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e59f-account-create-t5q2j" event={"ID":"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae","Type":"ContainerStarted","Data":"4252cad99b13bf30db5f9737bcbc367232d45b66ae8feb39f1ed0f5af8ef92eb"} Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.591385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e59f-account-create-t5q2j" event={"ID":"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae","Type":"ContainerStarted","Data":"95c237dade31fb7fb8f50f086a9bedba7b28d79a4a45b4e81a7ee9adfc91d695"} Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.595060 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0a19-account-create-gxxbb" event={"ID":"9f2ba678-6efb-4ec5-b163-da752d2edecb","Type":"ContainerStarted","Data":"e7bd5a1cf8ef04bc77baede9980f996b19cd411f7bd846c1b620c067ba0cdee1"} Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.595114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0a19-account-create-gxxbb" event={"ID":"9f2ba678-6efb-4ec5-b163-da752d2edecb","Type":"ContainerStarted","Data":"bc6abe9a21ee62df1fb65f2e7eaa4d90982b043964559e1dce19e37ea68c9ad9"} Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.597444 4824 generic.go:334] "Generic (PLEG): container finished" podID="8fc09e1d-e018-4086-b17e-f7a1ccf02951" containerID="54da7f04b24314fc7fa59d93fb6b64cceea5b86eb0b61c31358ef6fc00767738" exitCode=0 Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.597492 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mnhn5" event={"ID":"8fc09e1d-e018-4086-b17e-f7a1ccf02951","Type":"ContainerDied","Data":"54da7f04b24314fc7fa59d93fb6b64cceea5b86eb0b61c31358ef6fc00767738"} Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.601312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8dxxk" event={"ID":"6cbf4587-2c54-4220-a514-62b043882b5b","Type":"ContainerStarted","Data":"b73f7af5aa4e5276937e5032ce17b43502984db3fb1f39f12e321fca6cfd499c"} Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.601515 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8dxxk" event={"ID":"6cbf4587-2c54-4220-a514-62b043882b5b","Type":"ContainerStarted","Data":"85030ffa108bd8dfdd659393fa98fe1243e7410c81b0b4507d791fbb7579edc7"} Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.606184 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-f0d5-account-create-46mmw" podStartSLOduration=2.606169409 podStartE2EDuration="2.606169409s" podCreationTimestamp="2025-11-24 13:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:40.599557749 +0000 UTC m=+1282.239097059" watchObservedRunningTime="2025-11-24 13:35:40.606169409 +0000 UTC m=+1282.245708719" Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.656611 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-0a19-account-create-gxxbb" podStartSLOduration=2.656594788 podStartE2EDuration="2.656594788s" podCreationTimestamp="2025-11-24 13:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:40.650225224 +0000 UTC m=+1282.289764534" watchObservedRunningTime="2025-11-24 13:35:40.656594788 +0000 UTC m=+1282.296134098" Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.671550 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-e59f-account-create-t5q2j" podStartSLOduration=2.671536383 podStartE2EDuration="2.671536383s" podCreationTimestamp="2025-11-24 13:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:40.671104832 +0000 UTC m=+1282.310644142" watchObservedRunningTime="2025-11-24 13:35:40.671536383 +0000 UTC m=+1282.311075693" Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.688868 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-8dxxk" podStartSLOduration=2.688853329 podStartE2EDuration="2.688853329s" podCreationTimestamp="2025-11-24 13:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:35:40.685853601 +0000 UTC m=+1282.325392911" watchObservedRunningTime="2025-11-24 13:35:40.688853329 +0000 UTC m=+1282.328392639" Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.788316 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.788369 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.788411 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.789164 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7023e2fedbc8a159e146d80c57d0a65acc7f7defe0350732f194124c1f095b6e"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:35:40 crc kubenswrapper[4824]: I1124 13:35:40.789216 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://7023e2fedbc8a159e146d80c57d0a65acc7f7defe0350732f194124c1f095b6e" gracePeriod=600 Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.611600 4824 generic.go:334] "Generic (PLEG): container finished" podID="6cbf4587-2c54-4220-a514-62b043882b5b" containerID="b73f7af5aa4e5276937e5032ce17b43502984db3fb1f39f12e321fca6cfd499c" exitCode=0 Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.611938 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8dxxk" event={"ID":"6cbf4587-2c54-4220-a514-62b043882b5b","Type":"ContainerDied","Data":"b73f7af5aa4e5276937e5032ce17b43502984db3fb1f39f12e321fca6cfd499c"} Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.614476 4824 generic.go:334] "Generic (PLEG): container finished" podID="570330a6-c219-4a73-8731-b85ca896c82c" containerID="f1d7def3cc42ff723e398a77610016b05a96420dd0b60e3e0395650b2a042e20" exitCode=0 Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.614559 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f0d5-account-create-46mmw" event={"ID":"570330a6-c219-4a73-8731-b85ca896c82c","Type":"ContainerDied","Data":"f1d7def3cc42ff723e398a77610016b05a96420dd0b60e3e0395650b2a042e20"} Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.616005 4824 generic.go:334] "Generic (PLEG): container finished" podID="7e9dd708-d69f-4c23-bbe1-2f41c7f18fae" containerID="4252cad99b13bf30db5f9737bcbc367232d45b66ae8feb39f1ed0f5af8ef92eb" exitCode=0 Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.616063 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e59f-account-create-t5q2j" event={"ID":"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae","Type":"ContainerDied","Data":"4252cad99b13bf30db5f9737bcbc367232d45b66ae8feb39f1ed0f5af8ef92eb"} Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.617746 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f2ba678-6efb-4ec5-b163-da752d2edecb" containerID="e7bd5a1cf8ef04bc77baede9980f996b19cd411f7bd846c1b620c067ba0cdee1" exitCode=0 Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.617827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0a19-account-create-gxxbb" event={"ID":"9f2ba678-6efb-4ec5-b163-da752d2edecb","Type":"ContainerDied","Data":"e7bd5a1cf8ef04bc77baede9980f996b19cd411f7bd846c1b620c067ba0cdee1"} Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.620405 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="7023e2fedbc8a159e146d80c57d0a65acc7f7defe0350732f194124c1f095b6e" exitCode=0 Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.620436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"7023e2fedbc8a159e146d80c57d0a65acc7f7defe0350732f194124c1f095b6e"} Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.620573 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"b1981738b6eecfa8db6367aadef18ef52e9a453a009ef7bb2c198c0fe6a176d1"} Nov 24 13:35:41 crc kubenswrapper[4824]: I1124 13:35:41.620603 4824 scope.go:117] "RemoveContainer" containerID="62c6a3d914058b43356f45839d76c3c6a0c31d95c968d9b2fa78dd4dbda74ea8" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.138113 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.157478 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.174922 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc09e1d-e018-4086-b17e-f7a1ccf02951-operator-scripts\") pod \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\" (UID: \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\") " Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.175126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jx98\" (UniqueName: \"kubernetes.io/projected/8fc09e1d-e018-4086-b17e-f7a1ccf02951-kube-api-access-5jx98\") pod \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\" (UID: \"8fc09e1d-e018-4086-b17e-f7a1ccf02951\") " Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.175743 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fc09e1d-e018-4086-b17e-f7a1ccf02951-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8fc09e1d-e018-4086-b17e-f7a1ccf02951" (UID: "8fc09e1d-e018-4086-b17e-f7a1ccf02951"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.183463 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fc09e1d-e018-4086-b17e-f7a1ccf02951-kube-api-access-5jx98" (OuterVolumeSpecName: "kube-api-access-5jx98") pod "8fc09e1d-e018-4086-b17e-f7a1ccf02951" (UID: "8fc09e1d-e018-4086-b17e-f7a1ccf02951"). InnerVolumeSpecName "kube-api-access-5jx98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.276357 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2lr9\" (UniqueName: \"kubernetes.io/projected/f396378e-6a0d-4853-a622-3d5ca30d45fc-kube-api-access-b2lr9\") pod \"f396378e-6a0d-4853-a622-3d5ca30d45fc\" (UID: \"f396378e-6a0d-4853-a622-3d5ca30d45fc\") " Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.276517 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f396378e-6a0d-4853-a622-3d5ca30d45fc-operator-scripts\") pod \"f396378e-6a0d-4853-a622-3d5ca30d45fc\" (UID: \"f396378e-6a0d-4853-a622-3d5ca30d45fc\") " Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.276918 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jx98\" (UniqueName: \"kubernetes.io/projected/8fc09e1d-e018-4086-b17e-f7a1ccf02951-kube-api-access-5jx98\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.276934 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc09e1d-e018-4086-b17e-f7a1ccf02951-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.277230 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f396378e-6a0d-4853-a622-3d5ca30d45fc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f396378e-6a0d-4853-a622-3d5ca30d45fc" (UID: "f396378e-6a0d-4853-a622-3d5ca30d45fc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.283950 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f396378e-6a0d-4853-a622-3d5ca30d45fc-kube-api-access-b2lr9" (OuterVolumeSpecName: "kube-api-access-b2lr9") pod "f396378e-6a0d-4853-a622-3d5ca30d45fc" (UID: "f396378e-6a0d-4853-a622-3d5ca30d45fc"). InnerVolumeSpecName "kube-api-access-b2lr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.378301 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2lr9\" (UniqueName: \"kubernetes.io/projected/f396378e-6a0d-4853-a622-3d5ca30d45fc-kube-api-access-b2lr9\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.378336 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f396378e-6a0d-4853-a622-3d5ca30d45fc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.641724 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mnhn5" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.641781 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mnhn5" event={"ID":"8fc09e1d-e018-4086-b17e-f7a1ccf02951","Type":"ContainerDied","Data":"46bb48c4039626b4768b3567f43f875aec9a2f905b9013e4919fccbcf98ffd9f"} Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.642889 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46bb48c4039626b4768b3567f43f875aec9a2f905b9013e4919fccbcf98ffd9f" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.644442 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5h6sl" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.644881 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5h6sl" event={"ID":"f396378e-6a0d-4853-a622-3d5ca30d45fc","Type":"ContainerDied","Data":"6ab3f96d7e8d5df22a0f733667582627a49e0bbaec68d8f8a39f675fa53793a2"} Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.644902 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ab3f96d7e8d5df22a0f733667582627a49e0bbaec68d8f8a39f675fa53793a2" Nov 24 13:35:42 crc kubenswrapper[4824]: I1124 13:35:42.920001 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.012328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cbf4587-2c54-4220-a514-62b043882b5b-operator-scripts\") pod \"6cbf4587-2c54-4220-a514-62b043882b5b\" (UID: \"6cbf4587-2c54-4220-a514-62b043882b5b\") " Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.012406 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbvr5\" (UniqueName: \"kubernetes.io/projected/6cbf4587-2c54-4220-a514-62b043882b5b-kube-api-access-gbvr5\") pod \"6cbf4587-2c54-4220-a514-62b043882b5b\" (UID: \"6cbf4587-2c54-4220-a514-62b043882b5b\") " Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.023451 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbf4587-2c54-4220-a514-62b043882b5b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6cbf4587-2c54-4220-a514-62b043882b5b" (UID: "6cbf4587-2c54-4220-a514-62b043882b5b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.024357 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cbf4587-2c54-4220-a514-62b043882b5b-kube-api-access-gbvr5" (OuterVolumeSpecName: "kube-api-access-gbvr5") pod "6cbf4587-2c54-4220-a514-62b043882b5b" (UID: "6cbf4587-2c54-4220-a514-62b043882b5b"). InnerVolumeSpecName "kube-api-access-gbvr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.114500 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cbf4587-2c54-4220-a514-62b043882b5b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.114837 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbvr5\" (UniqueName: \"kubernetes.io/projected/6cbf4587-2c54-4220-a514-62b043882b5b-kube-api-access-gbvr5\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.451774 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.453338 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.453358 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.527612 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76skc\" (UniqueName: \"kubernetes.io/projected/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-kube-api-access-76skc\") pod \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\" (UID: \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\") " Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.527651 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570330a6-c219-4a73-8731-b85ca896c82c-operator-scripts\") pod \"570330a6-c219-4a73-8731-b85ca896c82c\" (UID: \"570330a6-c219-4a73-8731-b85ca896c82c\") " Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.527734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2ba678-6efb-4ec5-b163-da752d2edecb-operator-scripts\") pod \"9f2ba678-6efb-4ec5-b163-da752d2edecb\" (UID: \"9f2ba678-6efb-4ec5-b163-da752d2edecb\") " Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.527856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhfd8\" (UniqueName: \"kubernetes.io/projected/9f2ba678-6efb-4ec5-b163-da752d2edecb-kube-api-access-zhfd8\") pod \"9f2ba678-6efb-4ec5-b163-da752d2edecb\" (UID: \"9f2ba678-6efb-4ec5-b163-da752d2edecb\") " Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.527925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddc76\" (UniqueName: \"kubernetes.io/projected/570330a6-c219-4a73-8731-b85ca896c82c-kube-api-access-ddc76\") pod \"570330a6-c219-4a73-8731-b85ca896c82c\" (UID: \"570330a6-c219-4a73-8731-b85ca896c82c\") " Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.527975 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-operator-scripts\") pod \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\" (UID: \"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae\") " Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.528426 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/570330a6-c219-4a73-8731-b85ca896c82c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "570330a6-c219-4a73-8731-b85ca896c82c" (UID: "570330a6-c219-4a73-8731-b85ca896c82c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.528595 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f2ba678-6efb-4ec5-b163-da752d2edecb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9f2ba678-6efb-4ec5-b163-da752d2edecb" (UID: "9f2ba678-6efb-4ec5-b163-da752d2edecb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.528691 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e9dd708-d69f-4c23-bbe1-2f41c7f18fae" (UID: "7e9dd708-d69f-4c23-bbe1-2f41c7f18fae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.535785 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-kube-api-access-76skc" (OuterVolumeSpecName: "kube-api-access-76skc") pod "7e9dd708-d69f-4c23-bbe1-2f41c7f18fae" (UID: "7e9dd708-d69f-4c23-bbe1-2f41c7f18fae"). InnerVolumeSpecName "kube-api-access-76skc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.536180 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f2ba678-6efb-4ec5-b163-da752d2edecb-kube-api-access-zhfd8" (OuterVolumeSpecName: "kube-api-access-zhfd8") pod "9f2ba678-6efb-4ec5-b163-da752d2edecb" (UID: "9f2ba678-6efb-4ec5-b163-da752d2edecb"). InnerVolumeSpecName "kube-api-access-zhfd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.554021 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/570330a6-c219-4a73-8731-b85ca896c82c-kube-api-access-ddc76" (OuterVolumeSpecName: "kube-api-access-ddc76") pod "570330a6-c219-4a73-8731-b85ca896c82c" (UID: "570330a6-c219-4a73-8731-b85ca896c82c"). InnerVolumeSpecName "kube-api-access-ddc76". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.629444 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2ba678-6efb-4ec5-b163-da752d2edecb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.629479 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhfd8\" (UniqueName: \"kubernetes.io/projected/9f2ba678-6efb-4ec5-b163-da752d2edecb-kube-api-access-zhfd8\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.629490 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddc76\" (UniqueName: \"kubernetes.io/projected/570330a6-c219-4a73-8731-b85ca896c82c-kube-api-access-ddc76\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.629499 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.629507 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76skc\" (UniqueName: \"kubernetes.io/projected/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae-kube-api-access-76skc\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.629516 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/570330a6-c219-4a73-8731-b85ca896c82c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.653098 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e59f-account-create-t5q2j" event={"ID":"7e9dd708-d69f-4c23-bbe1-2f41c7f18fae","Type":"ContainerDied","Data":"95c237dade31fb7fb8f50f086a9bedba7b28d79a4a45b4e81a7ee9adfc91d695"} Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.653136 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95c237dade31fb7fb8f50f086a9bedba7b28d79a4a45b4e81a7ee9adfc91d695" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.653142 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e59f-account-create-t5q2j" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.658372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0a19-account-create-gxxbb" event={"ID":"9f2ba678-6efb-4ec5-b163-da752d2edecb","Type":"ContainerDied","Data":"bc6abe9a21ee62df1fb65f2e7eaa4d90982b043964559e1dce19e37ea68c9ad9"} Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.658404 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc6abe9a21ee62df1fb65f2e7eaa4d90982b043964559e1dce19e37ea68c9ad9" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.658447 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0a19-account-create-gxxbb" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.660091 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8dxxk" event={"ID":"6cbf4587-2c54-4220-a514-62b043882b5b","Type":"ContainerDied","Data":"85030ffa108bd8dfdd659393fa98fe1243e7410c81b0b4507d791fbb7579edc7"} Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.660118 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85030ffa108bd8dfdd659393fa98fe1243e7410c81b0b4507d791fbb7579edc7" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.660167 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8dxxk" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.662683 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f0d5-account-create-46mmw" Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.666622 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f0d5-account-create-46mmw" event={"ID":"570330a6-c219-4a73-8731-b85ca896c82c","Type":"ContainerDied","Data":"4d84a830f209a9251d4495a67ab029422fb60a97c707919bc451a4c9790cf5a3"} Nov 24 13:35:43 crc kubenswrapper[4824]: I1124 13:35:43.666661 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d84a830f209a9251d4495a67ab029422fb60a97c707919bc451a4c9790cf5a3" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.320349 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.342148 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-log-httpd\") pod \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.342280 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-combined-ca-bundle\") pod \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.342386 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-scripts\") pod \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.342441 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-sg-core-conf-yaml\") pod \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.342465 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hnbh\" (UniqueName: \"kubernetes.io/projected/ff67241a-4c14-4ebf-913a-69e8a5beb5af-kube-api-access-4hnbh\") pod \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.342519 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-config-data\") pod \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.342560 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-run-httpd\") pod \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\" (UID: \"ff67241a-4c14-4ebf-913a-69e8a5beb5af\") " Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.343501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ff67241a-4c14-4ebf-913a-69e8a5beb5af" (UID: "ff67241a-4c14-4ebf-913a-69e8a5beb5af"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.343639 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ff67241a-4c14-4ebf-913a-69e8a5beb5af" (UID: "ff67241a-4c14-4ebf-913a-69e8a5beb5af"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.352456 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff67241a-4c14-4ebf-913a-69e8a5beb5af-kube-api-access-4hnbh" (OuterVolumeSpecName: "kube-api-access-4hnbh") pod "ff67241a-4c14-4ebf-913a-69e8a5beb5af" (UID: "ff67241a-4c14-4ebf-913a-69e8a5beb5af"). InnerVolumeSpecName "kube-api-access-4hnbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.354276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-scripts" (OuterVolumeSpecName: "scripts") pod "ff67241a-4c14-4ebf-913a-69e8a5beb5af" (UID: "ff67241a-4c14-4ebf-913a-69e8a5beb5af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.426478 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ff67241a-4c14-4ebf-913a-69e8a5beb5af" (UID: "ff67241a-4c14-4ebf-913a-69e8a5beb5af"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.444913 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.444955 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.444964 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hnbh\" (UniqueName: \"kubernetes.io/projected/ff67241a-4c14-4ebf-913a-69e8a5beb5af-kube-api-access-4hnbh\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.444972 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.444980 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff67241a-4c14-4ebf-913a-69e8a5beb5af-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.462176 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff67241a-4c14-4ebf-913a-69e8a5beb5af" (UID: "ff67241a-4c14-4ebf-913a-69e8a5beb5af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.500785 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-config-data" (OuterVolumeSpecName: "config-data") pod "ff67241a-4c14-4ebf-913a-69e8a5beb5af" (UID: "ff67241a-4c14-4ebf-913a-69e8a5beb5af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.546191 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.546249 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff67241a-4c14-4ebf-913a-69e8a5beb5af-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.672526 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerID="0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa" exitCode=0 Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.672849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerDied","Data":"0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa"} Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.672876 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff67241a-4c14-4ebf-913a-69e8a5beb5af","Type":"ContainerDied","Data":"0ad7cf68605235b6d92551147f520938342b5ce3877070bbca07054be7f7b563"} Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.672894 4824 scope.go:117] "RemoveContainer" containerID="53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.673007 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.702455 4824 scope.go:117] "RemoveContainer" containerID="ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.716087 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.728250 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.729417 4824 scope.go:117] "RemoveContainer" containerID="7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.747696 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.751840 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="ceilometer-central-agent" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.751869 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="ceilometer-central-agent" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.751881 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbf4587-2c54-4220-a514-62b043882b5b" containerName="mariadb-database-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.751887 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbf4587-2c54-4220-a514-62b043882b5b" containerName="mariadb-database-create" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.751909 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fc09e1d-e018-4086-b17e-f7a1ccf02951" containerName="mariadb-database-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.751917 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fc09e1d-e018-4086-b17e-f7a1ccf02951" containerName="mariadb-database-create" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.751930 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="sg-core" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.751935 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="sg-core" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.751956 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="570330a6-c219-4a73-8731-b85ca896c82c" containerName="mariadb-account-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.751966 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="570330a6-c219-4a73-8731-b85ca896c82c" containerName="mariadb-account-create" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.751981 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="ceilometer-notification-agent" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.751987 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="ceilometer-notification-agent" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.752001 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9dd708-d69f-4c23-bbe1-2f41c7f18fae" containerName="mariadb-account-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752007 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9dd708-d69f-4c23-bbe1-2f41c7f18fae" containerName="mariadb-account-create" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.752020 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2ba678-6efb-4ec5-b163-da752d2edecb" containerName="mariadb-account-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752026 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2ba678-6efb-4ec5-b163-da752d2edecb" containerName="mariadb-account-create" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.752040 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f396378e-6a0d-4853-a622-3d5ca30d45fc" containerName="mariadb-database-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752046 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f396378e-6a0d-4853-a622-3d5ca30d45fc" containerName="mariadb-database-create" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.752058 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="proxy-httpd" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752063 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="proxy-httpd" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752291 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cbf4587-2c54-4220-a514-62b043882b5b" containerName="mariadb-database-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752310 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="570330a6-c219-4a73-8731-b85ca896c82c" containerName="mariadb-account-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752321 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="sg-core" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752328 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="proxy-httpd" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752353 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9dd708-d69f-4c23-bbe1-2f41c7f18fae" containerName="mariadb-account-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752366 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f2ba678-6efb-4ec5-b163-da752d2edecb" containerName="mariadb-account-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752381 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="ceilometer-notification-agent" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752396 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f396378e-6a0d-4853-a622-3d5ca30d45fc" containerName="mariadb-database-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752411 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" containerName="ceilometer-central-agent" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.752419 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fc09e1d-e018-4086-b17e-f7a1ccf02951" containerName="mariadb-database-create" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.754088 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.756539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.758261 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.765114 4824 scope.go:117] "RemoveContainer" containerID="0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.766665 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.801113 4824 scope.go:117] "RemoveContainer" containerID="53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.806010 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04\": container with ID starting with 53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04 not found: ID does not exist" containerID="53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.806043 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04"} err="failed to get container status \"53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04\": rpc error: code = NotFound desc = could not find container \"53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04\": container with ID starting with 53a4eedc85097c1da93fa9e53736a533bbfd740701d3289db6d0f71185d6ef04 not found: ID does not exist" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.806064 4824 scope.go:117] "RemoveContainer" containerID="ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.806629 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f\": container with ID starting with ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f not found: ID does not exist" containerID="ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.806672 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f"} err="failed to get container status \"ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f\": rpc error: code = NotFound desc = could not find container \"ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f\": container with ID starting with ede929b3a132651ba061072c509be8db440cd8a21ad55f734a164f8882197d5f not found: ID does not exist" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.806700 4824 scope.go:117] "RemoveContainer" containerID="7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.807244 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d\": container with ID starting with 7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d not found: ID does not exist" containerID="7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.807270 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d"} err="failed to get container status \"7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d\": rpc error: code = NotFound desc = could not find container \"7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d\": container with ID starting with 7c65b21fed5336dc6a35d9da2d132a6074298470c2d82b0fa0c637f8a80b117d not found: ID does not exist" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.807286 4824 scope.go:117] "RemoveContainer" containerID="0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa" Nov 24 13:35:44 crc kubenswrapper[4824]: E1124 13:35:44.807520 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa\": container with ID starting with 0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa not found: ID does not exist" containerID="0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.807553 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa"} err="failed to get container status \"0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa\": rpc error: code = NotFound desc = could not find container \"0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa\": container with ID starting with 0864f7e5fc1eaa2b3bdd77795fe9aa69f02dc8b1cf28c8c55d4bac489ac3d4aa not found: ID does not exist" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.851346 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-run-httpd\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.851396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-scripts\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.851423 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.851459 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-config-data\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.851493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-log-httpd\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.851661 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.851723 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj9t8\" (UniqueName: \"kubernetes.io/projected/28e1651b-29a2-498f-a17e-7d8e50624132-kube-api-access-lj9t8\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.953175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-run-httpd\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.953228 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-scripts\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.953249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.953287 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-config-data\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.953307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-log-httpd\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.953347 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.953372 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj9t8\" (UniqueName: \"kubernetes.io/projected/28e1651b-29a2-498f-a17e-7d8e50624132-kube-api-access-lj9t8\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.953775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-run-httpd\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.954189 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-log-httpd\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.958683 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.959530 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-config-data\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.960444 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.964385 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-scripts\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:44 crc kubenswrapper[4824]: I1124 13:35:44.970177 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj9t8\" (UniqueName: \"kubernetes.io/projected/28e1651b-29a2-498f-a17e-7d8e50624132-kube-api-access-lj9t8\") pod \"ceilometer-0\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " pod="openstack/ceilometer-0" Nov 24 13:35:45 crc kubenswrapper[4824]: I1124 13:35:45.020664 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff67241a-4c14-4ebf-913a-69e8a5beb5af" path="/var/lib/kubelet/pods/ff67241a-4c14-4ebf-913a-69e8a5beb5af/volumes" Nov 24 13:35:45 crc kubenswrapper[4824]: I1124 13:35:45.077986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:35:45 crc kubenswrapper[4824]: I1124 13:35:45.523243 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:35:45 crc kubenswrapper[4824]: I1124 13:35:45.688004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerStarted","Data":"eb097af6c1bcfa311dbd9773f57b830d4c86e9b3d7d46380659adb79bee8bf53"} Nov 24 13:35:46 crc kubenswrapper[4824]: I1124 13:35:46.701549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerStarted","Data":"08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626"} Nov 24 13:35:47 crc kubenswrapper[4824]: I1124 13:35:47.711509 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerStarted","Data":"742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2"} Nov 24 13:35:47 crc kubenswrapper[4824]: I1124 13:35:47.712196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerStarted","Data":"f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b"} Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.655881 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7xbd2"] Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.656989 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.659008 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.659508 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.660120 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-x47ph" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.686034 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7xbd2"] Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.736060 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.736166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-scripts\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.736354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnwjr\" (UniqueName: \"kubernetes.io/projected/83a86b9f-f865-4d11-8d95-ac103ea18ca7-kube-api-access-pnwjr\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.736375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-config-data\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.838632 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnwjr\" (UniqueName: \"kubernetes.io/projected/83a86b9f-f865-4d11-8d95-ac103ea18ca7-kube-api-access-pnwjr\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.838699 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-config-data\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.838780 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.838864 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-scripts\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.845739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.856368 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-scripts\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.856882 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnwjr\" (UniqueName: \"kubernetes.io/projected/83a86b9f-f865-4d11-8d95-ac103ea18ca7-kube-api-access-pnwjr\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.866639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-config-data\") pod \"nova-cell0-conductor-db-sync-7xbd2\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:48 crc kubenswrapper[4824]: I1124 13:35:48.984910 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:35:49 crc kubenswrapper[4824]: I1124 13:35:49.544203 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7xbd2"] Nov 24 13:35:49 crc kubenswrapper[4824]: I1124 13:35:49.733360 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerStarted","Data":"8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090"} Nov 24 13:35:49 crc kubenswrapper[4824]: I1124 13:35:49.733694 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:35:49 crc kubenswrapper[4824]: I1124 13:35:49.734343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7xbd2" event={"ID":"83a86b9f-f865-4d11-8d95-ac103ea18ca7","Type":"ContainerStarted","Data":"92440881bfb16865506de15b3160dd89de4f0d13aa4677e661014a319585a03e"} Nov 24 13:35:49 crc kubenswrapper[4824]: I1124 13:35:49.774310 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.303586671 podStartE2EDuration="5.774287244s" podCreationTimestamp="2025-11-24 13:35:44 +0000 UTC" firstStartedPulling="2025-11-24 13:35:45.537589796 +0000 UTC m=+1287.177129106" lastFinishedPulling="2025-11-24 13:35:49.008290369 +0000 UTC m=+1290.647829679" observedRunningTime="2025-11-24 13:35:49.764725488 +0000 UTC m=+1291.404264798" watchObservedRunningTime="2025-11-24 13:35:49.774287244 +0000 UTC m=+1291.413826554" Nov 24 13:35:52 crc kubenswrapper[4824]: I1124 13:35:52.595721 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:35:52 crc kubenswrapper[4824]: I1124 13:35:52.597573 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:35:54 crc kubenswrapper[4824]: I1124 13:35:54.293878 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-959bb6dd8-8pcrj" Nov 24 13:35:54 crc kubenswrapper[4824]: I1124 13:35:54.358901 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c96f4646d-wsbgv"] Nov 24 13:35:54 crc kubenswrapper[4824]: I1124 13:35:54.359132 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon-log" containerID="cri-o://8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933" gracePeriod=30 Nov 24 13:35:54 crc kubenswrapper[4824]: I1124 13:35:54.359866 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" containerID="cri-o://1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f" gracePeriod=30 Nov 24 13:35:54 crc kubenswrapper[4824]: I1124 13:35:54.398940 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Nov 24 13:35:57 crc kubenswrapper[4824]: I1124 13:35:57.533106 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:51450->10.217.0.147:8443: read: connection reset by peer" Nov 24 13:35:57 crc kubenswrapper[4824]: I1124 13:35:57.814859 4824 generic.go:334] "Generic (PLEG): container finished" podID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerID="1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f" exitCode=0 Nov 24 13:35:57 crc kubenswrapper[4824]: I1124 13:35:57.814911 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c96f4646d-wsbgv" event={"ID":"bc2d2fae-b656-464e-b1a8-89b4c5160792","Type":"ContainerDied","Data":"1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f"} Nov 24 13:35:57 crc kubenswrapper[4824]: I1124 13:35:57.815251 4824 scope.go:117] "RemoveContainer" containerID="6ce765b74f266b883a095df6fbfd4992669f9774df16a85346b8570f4fc1ede0" Nov 24 13:35:59 crc kubenswrapper[4824]: I1124 13:35:59.467464 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 24 13:35:59 crc kubenswrapper[4824]: I1124 13:35:59.842271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7xbd2" event={"ID":"83a86b9f-f865-4d11-8d95-ac103ea18ca7","Type":"ContainerStarted","Data":"83ad15175826da0d74b1ff9e2c38c2c74235e4e15b71addb9712beac243cd8ca"} Nov 24 13:35:59 crc kubenswrapper[4824]: I1124 13:35:59.867274 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7xbd2" podStartSLOduration=3.000294324 podStartE2EDuration="11.867253507s" podCreationTimestamp="2025-11-24 13:35:48 +0000 UTC" firstStartedPulling="2025-11-24 13:35:49.553907849 +0000 UTC m=+1291.193447159" lastFinishedPulling="2025-11-24 13:35:58.420867022 +0000 UTC m=+1300.060406342" observedRunningTime="2025-11-24 13:35:59.856480693 +0000 UTC m=+1301.496020013" watchObservedRunningTime="2025-11-24 13:35:59.867253507 +0000 UTC m=+1301.506792827" Nov 24 13:36:09 crc kubenswrapper[4824]: I1124 13:36:09.467328 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 24 13:36:12 crc kubenswrapper[4824]: I1124 13:36:12.000030 4824 generic.go:334] "Generic (PLEG): container finished" podID="83a86b9f-f865-4d11-8d95-ac103ea18ca7" containerID="83ad15175826da0d74b1ff9e2c38c2c74235e4e15b71addb9712beac243cd8ca" exitCode=0 Nov 24 13:36:12 crc kubenswrapper[4824]: I1124 13:36:12.000153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7xbd2" event={"ID":"83a86b9f-f865-4d11-8d95-ac103ea18ca7","Type":"ContainerDied","Data":"83ad15175826da0d74b1ff9e2c38c2c74235e4e15b71addb9712beac243cd8ca"} Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.436881 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.550171 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnwjr\" (UniqueName: \"kubernetes.io/projected/83a86b9f-f865-4d11-8d95-ac103ea18ca7-kube-api-access-pnwjr\") pod \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.550353 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-combined-ca-bundle\") pod \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.550403 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-config-data\") pod \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.550448 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-scripts\") pod \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\" (UID: \"83a86b9f-f865-4d11-8d95-ac103ea18ca7\") " Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.563615 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-scripts" (OuterVolumeSpecName: "scripts") pod "83a86b9f-f865-4d11-8d95-ac103ea18ca7" (UID: "83a86b9f-f865-4d11-8d95-ac103ea18ca7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.569227 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a86b9f-f865-4d11-8d95-ac103ea18ca7-kube-api-access-pnwjr" (OuterVolumeSpecName: "kube-api-access-pnwjr") pod "83a86b9f-f865-4d11-8d95-ac103ea18ca7" (UID: "83a86b9f-f865-4d11-8d95-ac103ea18ca7"). InnerVolumeSpecName "kube-api-access-pnwjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.578177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83a86b9f-f865-4d11-8d95-ac103ea18ca7" (UID: "83a86b9f-f865-4d11-8d95-ac103ea18ca7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.587699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-config-data" (OuterVolumeSpecName: "config-data") pod "83a86b9f-f865-4d11-8d95-ac103ea18ca7" (UID: "83a86b9f-f865-4d11-8d95-ac103ea18ca7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.665997 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.666050 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnwjr\" (UniqueName: \"kubernetes.io/projected/83a86b9f-f865-4d11-8d95-ac103ea18ca7-kube-api-access-pnwjr\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.666067 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:13 crc kubenswrapper[4824]: I1124 13:36:13.666079 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a86b9f-f865-4d11-8d95-ac103ea18ca7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.022574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7xbd2" event={"ID":"83a86b9f-f865-4d11-8d95-ac103ea18ca7","Type":"ContainerDied","Data":"92440881bfb16865506de15b3160dd89de4f0d13aa4677e661014a319585a03e"} Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.022992 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92440881bfb16865506de15b3160dd89de4f0d13aa4677e661014a319585a03e" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.023088 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7xbd2" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.170135 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 13:36:14 crc kubenswrapper[4824]: E1124 13:36:14.170547 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a86b9f-f865-4d11-8d95-ac103ea18ca7" containerName="nova-cell0-conductor-db-sync" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.170568 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a86b9f-f865-4d11-8d95-ac103ea18ca7" containerName="nova-cell0-conductor-db-sync" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.170850 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a86b9f-f865-4d11-8d95-ac103ea18ca7" containerName="nova-cell0-conductor-db-sync" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.171563 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.173984 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.174273 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-x47ph" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.199854 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.276658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5819ad8-0aa5-4d68-831e-09c741267fd5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e5819ad8-0aa5-4d68-831e-09c741267fd5\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.276770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr45c\" (UniqueName: \"kubernetes.io/projected/e5819ad8-0aa5-4d68-831e-09c741267fd5-kube-api-access-wr45c\") pod \"nova-cell0-conductor-0\" (UID: \"e5819ad8-0aa5-4d68-831e-09c741267fd5\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.276869 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5819ad8-0aa5-4d68-831e-09c741267fd5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e5819ad8-0aa5-4d68-831e-09c741267fd5\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.378144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5819ad8-0aa5-4d68-831e-09c741267fd5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e5819ad8-0aa5-4d68-831e-09c741267fd5\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.378408 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr45c\" (UniqueName: \"kubernetes.io/projected/e5819ad8-0aa5-4d68-831e-09c741267fd5-kube-api-access-wr45c\") pod \"nova-cell0-conductor-0\" (UID: \"e5819ad8-0aa5-4d68-831e-09c741267fd5\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.378503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5819ad8-0aa5-4d68-831e-09c741267fd5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e5819ad8-0aa5-4d68-831e-09c741267fd5\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.395298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5819ad8-0aa5-4d68-831e-09c741267fd5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e5819ad8-0aa5-4d68-831e-09c741267fd5\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.395576 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5819ad8-0aa5-4d68-831e-09c741267fd5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e5819ad8-0aa5-4d68-831e-09c741267fd5\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.407849 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr45c\" (UniqueName: \"kubernetes.io/projected/e5819ad8-0aa5-4d68-831e-09c741267fd5-kube-api-access-wr45c\") pod \"nova-cell0-conductor-0\" (UID: \"e5819ad8-0aa5-4d68-831e-09c741267fd5\") " pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.528821 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:14 crc kubenswrapper[4824]: I1124 13:36:14.983473 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 13:36:15 crc kubenswrapper[4824]: I1124 13:36:15.041955 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e5819ad8-0aa5-4d68-831e-09c741267fd5","Type":"ContainerStarted","Data":"e79be82c04349e5c776cb36619f52989797d4356ba4bf53bed350e6127b50795"} Nov 24 13:36:15 crc kubenswrapper[4824]: I1124 13:36:15.091426 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 13:36:16 crc kubenswrapper[4824]: I1124 13:36:16.053715 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e5819ad8-0aa5-4d68-831e-09c741267fd5","Type":"ContainerStarted","Data":"bb6a39afe09aea72954f2a601cfb089daaafd3422f13112220c9571b7a6041ab"} Nov 24 13:36:16 crc kubenswrapper[4824]: I1124 13:36:16.082029 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.082008039 podStartE2EDuration="2.082008039s" podCreationTimestamp="2025-11-24 13:36:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:16.079511555 +0000 UTC m=+1317.719050865" watchObservedRunningTime="2025-11-24 13:36:16.082008039 +0000 UTC m=+1317.721547349" Nov 24 13:36:17 crc kubenswrapper[4824]: I1124 13:36:17.060592 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:18 crc kubenswrapper[4824]: I1124 13:36:18.911134 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:36:18 crc kubenswrapper[4824]: I1124 13:36:18.911606 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="25fd29f3-f0aa-4062-9f63-19fe3218583c" containerName="kube-state-metrics" containerID="cri-o://e31406a580667045cf0323d2386cd7e782e8611712a0e39a616c2fb768c4114e" gracePeriod=30 Nov 24 13:36:19 crc kubenswrapper[4824]: I1124 13:36:19.080690 4824 generic.go:334] "Generic (PLEG): container finished" podID="25fd29f3-f0aa-4062-9f63-19fe3218583c" containerID="e31406a580667045cf0323d2386cd7e782e8611712a0e39a616c2fb768c4114e" exitCode=2 Nov 24 13:36:19 crc kubenswrapper[4824]: I1124 13:36:19.080736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25fd29f3-f0aa-4062-9f63-19fe3218583c","Type":"ContainerDied","Data":"e31406a580667045cf0323d2386cd7e782e8611712a0e39a616c2fb768c4114e"} Nov 24 13:36:19 crc kubenswrapper[4824]: I1124 13:36:19.374285 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:36:19 crc kubenswrapper[4824]: I1124 13:36:19.466613 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c96f4646d-wsbgv" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 24 13:36:19 crc kubenswrapper[4824]: I1124 13:36:19.466801 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmr6v\" (UniqueName: \"kubernetes.io/projected/25fd29f3-f0aa-4062-9f63-19fe3218583c-kube-api-access-cmr6v\") pod \"25fd29f3-f0aa-4062-9f63-19fe3218583c\" (UID: \"25fd29f3-f0aa-4062-9f63-19fe3218583c\") " Nov 24 13:36:19 crc kubenswrapper[4824]: I1124 13:36:19.483405 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25fd29f3-f0aa-4062-9f63-19fe3218583c-kube-api-access-cmr6v" (OuterVolumeSpecName: "kube-api-access-cmr6v") pod "25fd29f3-f0aa-4062-9f63-19fe3218583c" (UID: "25fd29f3-f0aa-4062-9f63-19fe3218583c"). InnerVolumeSpecName "kube-api-access-cmr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:19 crc kubenswrapper[4824]: I1124 13:36:19.568658 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmr6v\" (UniqueName: \"kubernetes.io/projected/25fd29f3-f0aa-4062-9f63-19fe3218583c-kube-api-access-cmr6v\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.090848 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25fd29f3-f0aa-4062-9f63-19fe3218583c","Type":"ContainerDied","Data":"1f50a1af5f6c5e4f9a404b0f7360055587e390478b40d0e82a59ebc9c22a6f0e"} Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.091718 4824 scope.go:117] "RemoveContainer" containerID="e31406a580667045cf0323d2386cd7e782e8611712a0e39a616c2fb768c4114e" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.090893 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.130617 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.147292 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.158148 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:36:20 crc kubenswrapper[4824]: E1124 13:36:20.158576 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25fd29f3-f0aa-4062-9f63-19fe3218583c" containerName="kube-state-metrics" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.158592 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="25fd29f3-f0aa-4062-9f63-19fe3218583c" containerName="kube-state-metrics" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.158790 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="25fd29f3-f0aa-4062-9f63-19fe3218583c" containerName="kube-state-metrics" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.159379 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.161718 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.161903 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.170047 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.281965 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a804c88-d42d-43ab-b316-d196ba26f5d1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.282036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk9hh\" (UniqueName: \"kubernetes.io/projected/8a804c88-d42d-43ab-b316-d196ba26f5d1-kube-api-access-tk9hh\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.282195 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8a804c88-d42d-43ab-b316-d196ba26f5d1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.282303 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a804c88-d42d-43ab-b316-d196ba26f5d1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.384605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a804c88-d42d-43ab-b316-d196ba26f5d1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.384739 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk9hh\" (UniqueName: \"kubernetes.io/projected/8a804c88-d42d-43ab-b316-d196ba26f5d1-kube-api-access-tk9hh\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.384859 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8a804c88-d42d-43ab-b316-d196ba26f5d1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.384934 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a804c88-d42d-43ab-b316-d196ba26f5d1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.390268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a804c88-d42d-43ab-b316-d196ba26f5d1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.390553 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a804c88-d42d-43ab-b316-d196ba26f5d1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.416961 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8a804c88-d42d-43ab-b316-d196ba26f5d1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.417568 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk9hh\" (UniqueName: \"kubernetes.io/projected/8a804c88-d42d-43ab-b316-d196ba26f5d1-kube-api-access-tk9hh\") pod \"kube-state-metrics-0\" (UID: \"8a804c88-d42d-43ab-b316-d196ba26f5d1\") " pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.490656 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.851823 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.852378 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="ceilometer-central-agent" containerID="cri-o://08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626" gracePeriod=30 Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.852843 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="sg-core" containerID="cri-o://742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2" gracePeriod=30 Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.852857 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="proxy-httpd" containerID="cri-o://8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090" gracePeriod=30 Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.853905 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="ceilometer-notification-agent" containerID="cri-o://f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b" gracePeriod=30 Nov 24 13:36:20 crc kubenswrapper[4824]: I1124 13:36:20.953641 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 13:36:20 crc kubenswrapper[4824]: W1124 13:36:20.956699 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a804c88_d42d_43ab_b316_d196ba26f5d1.slice/crio-6066e6992d0421a1d88b4af20162357eaee11bac39327cd6888a96b96d3508a2 WatchSource:0}: Error finding container 6066e6992d0421a1d88b4af20162357eaee11bac39327cd6888a96b96d3508a2: Status 404 returned error can't find the container with id 6066e6992d0421a1d88b4af20162357eaee11bac39327cd6888a96b96d3508a2 Nov 24 13:36:21 crc kubenswrapper[4824]: I1124 13:36:21.019785 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25fd29f3-f0aa-4062-9f63-19fe3218583c" path="/var/lib/kubelet/pods/25fd29f3-f0aa-4062-9f63-19fe3218583c/volumes" Nov 24 13:36:21 crc kubenswrapper[4824]: I1124 13:36:21.105507 4824 generic.go:334] "Generic (PLEG): container finished" podID="28e1651b-29a2-498f-a17e-7d8e50624132" containerID="8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090" exitCode=0 Nov 24 13:36:21 crc kubenswrapper[4824]: I1124 13:36:21.105837 4824 generic.go:334] "Generic (PLEG): container finished" podID="28e1651b-29a2-498f-a17e-7d8e50624132" containerID="742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2" exitCode=2 Nov 24 13:36:21 crc kubenswrapper[4824]: I1124 13:36:21.105575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerDied","Data":"8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090"} Nov 24 13:36:21 crc kubenswrapper[4824]: I1124 13:36:21.105942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerDied","Data":"742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2"} Nov 24 13:36:21 crc kubenswrapper[4824]: I1124 13:36:21.107127 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8a804c88-d42d-43ab-b316-d196ba26f5d1","Type":"ContainerStarted","Data":"6066e6992d0421a1d88b4af20162357eaee11bac39327cd6888a96b96d3508a2"} Nov 24 13:36:22 crc kubenswrapper[4824]: I1124 13:36:22.120025 4824 generic.go:334] "Generic (PLEG): container finished" podID="28e1651b-29a2-498f-a17e-7d8e50624132" containerID="08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626" exitCode=0 Nov 24 13:36:22 crc kubenswrapper[4824]: I1124 13:36:22.120138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerDied","Data":"08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626"} Nov 24 13:36:22 crc kubenswrapper[4824]: I1124 13:36:22.122308 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8a804c88-d42d-43ab-b316-d196ba26f5d1","Type":"ContainerStarted","Data":"601e887fe4f42d14b6733b09054f13ec2840a625e34af8f88b6b820fc2d5bc0d"} Nov 24 13:36:22 crc kubenswrapper[4824]: I1124 13:36:22.122434 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 13:36:22 crc kubenswrapper[4824]: I1124 13:36:22.139547 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.713201314 podStartE2EDuration="2.13953201s" podCreationTimestamp="2025-11-24 13:36:20 +0000 UTC" firstStartedPulling="2025-11-24 13:36:20.959884844 +0000 UTC m=+1322.599424154" lastFinishedPulling="2025-11-24 13:36:21.38621555 +0000 UTC m=+1323.025754850" observedRunningTime="2025-11-24 13:36:22.138455247 +0000 UTC m=+1323.777994597" watchObservedRunningTime="2025-11-24 13:36:22.13953201 +0000 UTC m=+1323.779071320" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.582619 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.821872 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.827104 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881479 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-secret-key\") pod \"bc2d2fae-b656-464e-b1a8-89b4c5160792\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881554 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-config-data\") pod \"28e1651b-29a2-498f-a17e-7d8e50624132\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881583 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-scripts\") pod \"bc2d2fae-b656-464e-b1a8-89b4c5160792\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881604 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-run-httpd\") pod \"28e1651b-29a2-498f-a17e-7d8e50624132\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881643 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d2fae-b656-464e-b1a8-89b4c5160792-logs\") pod \"bc2d2fae-b656-464e-b1a8-89b4c5160792\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881709 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-scripts\") pod \"28e1651b-29a2-498f-a17e-7d8e50624132\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-sg-core-conf-yaml\") pod \"28e1651b-29a2-498f-a17e-7d8e50624132\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881770 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj9t8\" (UniqueName: \"kubernetes.io/projected/28e1651b-29a2-498f-a17e-7d8e50624132-kube-api-access-lj9t8\") pod \"28e1651b-29a2-498f-a17e-7d8e50624132\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-combined-ca-bundle\") pod \"bc2d2fae-b656-464e-b1a8-89b4c5160792\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881855 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqp5f\" (UniqueName: \"kubernetes.io/projected/bc2d2fae-b656-464e-b1a8-89b4c5160792-kube-api-access-mqp5f\") pod \"bc2d2fae-b656-464e-b1a8-89b4c5160792\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-log-httpd\") pod \"28e1651b-29a2-498f-a17e-7d8e50624132\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881905 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-tls-certs\") pod \"bc2d2fae-b656-464e-b1a8-89b4c5160792\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881940 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-config-data\") pod \"bc2d2fae-b656-464e-b1a8-89b4c5160792\" (UID: \"bc2d2fae-b656-464e-b1a8-89b4c5160792\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.881976 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-combined-ca-bundle\") pod \"28e1651b-29a2-498f-a17e-7d8e50624132\" (UID: \"28e1651b-29a2-498f-a17e-7d8e50624132\") " Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.886525 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "28e1651b-29a2-498f-a17e-7d8e50624132" (UID: "28e1651b-29a2-498f-a17e-7d8e50624132"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.890178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "28e1651b-29a2-498f-a17e-7d8e50624132" (UID: "28e1651b-29a2-498f-a17e-7d8e50624132"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.890516 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2d2fae-b656-464e-b1a8-89b4c5160792-logs" (OuterVolumeSpecName: "logs") pod "bc2d2fae-b656-464e-b1a8-89b4c5160792" (UID: "bc2d2fae-b656-464e-b1a8-89b4c5160792"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.891084 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bc2d2fae-b656-464e-b1a8-89b4c5160792" (UID: "bc2d2fae-b656-464e-b1a8-89b4c5160792"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.907608 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-scripts" (OuterVolumeSpecName: "scripts") pod "28e1651b-29a2-498f-a17e-7d8e50624132" (UID: "28e1651b-29a2-498f-a17e-7d8e50624132"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.909213 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc2d2fae-b656-464e-b1a8-89b4c5160792-kube-api-access-mqp5f" (OuterVolumeSpecName: "kube-api-access-mqp5f") pod "bc2d2fae-b656-464e-b1a8-89b4c5160792" (UID: "bc2d2fae-b656-464e-b1a8-89b4c5160792"). InnerVolumeSpecName "kube-api-access-mqp5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.911016 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28e1651b-29a2-498f-a17e-7d8e50624132-kube-api-access-lj9t8" (OuterVolumeSpecName: "kube-api-access-lj9t8") pod "28e1651b-29a2-498f-a17e-7d8e50624132" (UID: "28e1651b-29a2-498f-a17e-7d8e50624132"). InnerVolumeSpecName "kube-api-access-lj9t8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.938232 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-config-data" (OuterVolumeSpecName: "config-data") pod "bc2d2fae-b656-464e-b1a8-89b4c5160792" (UID: "bc2d2fae-b656-464e-b1a8-89b4c5160792"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.961821 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "28e1651b-29a2-498f-a17e-7d8e50624132" (UID: "28e1651b-29a2-498f-a17e-7d8e50624132"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.963390 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc2d2fae-b656-464e-b1a8-89b4c5160792" (UID: "bc2d2fae-b656-464e-b1a8-89b4c5160792"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.963486 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-scripts" (OuterVolumeSpecName: "scripts") pod "bc2d2fae-b656-464e-b1a8-89b4c5160792" (UID: "bc2d2fae-b656-464e-b1a8-89b4c5160792"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986045 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986087 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986099 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986109 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d2fae-b656-464e-b1a8-89b4c5160792-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986120 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986131 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986142 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj9t8\" (UniqueName: \"kubernetes.io/projected/28e1651b-29a2-498f-a17e-7d8e50624132-kube-api-access-lj9t8\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986153 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986166 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqp5f\" (UniqueName: \"kubernetes.io/projected/bc2d2fae-b656-464e-b1a8-89b4c5160792-kube-api-access-mqp5f\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986176 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28e1651b-29a2-498f-a17e-7d8e50624132-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.986186 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc2d2fae-b656-464e-b1a8-89b4c5160792-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:24 crc kubenswrapper[4824]: I1124 13:36:24.991440 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "bc2d2fae-b656-464e-b1a8-89b4c5160792" (UID: "bc2d2fae-b656-464e-b1a8-89b4c5160792"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.005795 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28e1651b-29a2-498f-a17e-7d8e50624132" (UID: "28e1651b-29a2-498f-a17e-7d8e50624132"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.040093 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-config-data" (OuterVolumeSpecName: "config-data") pod "28e1651b-29a2-498f-a17e-7d8e50624132" (UID: "28e1651b-29a2-498f-a17e-7d8e50624132"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.088014 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.088047 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28e1651b-29a2-498f-a17e-7d8e50624132-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.088058 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2d2fae-b656-464e-b1a8-89b4c5160792-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.161845 4824 generic.go:334] "Generic (PLEG): container finished" podID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerID="8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933" exitCode=137 Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.162028 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c96f4646d-wsbgv" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.162205 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c96f4646d-wsbgv" event={"ID":"bc2d2fae-b656-464e-b1a8-89b4c5160792","Type":"ContainerDied","Data":"8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933"} Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.162251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c96f4646d-wsbgv" event={"ID":"bc2d2fae-b656-464e-b1a8-89b4c5160792","Type":"ContainerDied","Data":"95426b1d4e810601f0275b28c86b9c913b291bd7d431bf89c1f4441239ce5cbd"} Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.162270 4824 scope.go:117] "RemoveContainer" containerID="1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.167914 4824 generic.go:334] "Generic (PLEG): container finished" podID="28e1651b-29a2-498f-a17e-7d8e50624132" containerID="f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b" exitCode=0 Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.167976 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.167954 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerDied","Data":"f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b"} Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.168714 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"28e1651b-29a2-498f-a17e-7d8e50624132","Type":"ContainerDied","Data":"eb097af6c1bcfa311dbd9773f57b830d4c86e9b3d7d46380659adb79bee8bf53"} Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.194363 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c96f4646d-wsbgv"] Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.203145 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c96f4646d-wsbgv"] Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.229357 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.236075 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256075 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.256443 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="proxy-httpd" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256460 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="proxy-httpd" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.256475 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="ceilometer-central-agent" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256482 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="ceilometer-central-agent" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.256498 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon-log" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256503 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon-log" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.256521 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256529 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.256544 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256551 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.256560 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="sg-core" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256566 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="sg-core" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.256581 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="ceilometer-notification-agent" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256587 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="ceilometer-notification-agent" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256780 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="proxy-httpd" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256792 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon-log" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256826 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="ceilometer-notification-agent" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.256993 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="ceilometer-central-agent" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.257011 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" containerName="sg-core" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.257025 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.257038 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" containerName="horizon" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.258834 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.263022 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.263269 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.263477 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.287447 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.290024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-run-httpd\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.290306 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-log-httpd\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.290371 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.290411 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.290509 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.290538 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbxsq\" (UniqueName: \"kubernetes.io/projected/d83e37fd-645e-4258-a441-a9a578844485-kube-api-access-rbxsq\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.290657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-scripts\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.290706 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-config-data\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.365618 4824 scope.go:117] "RemoveContainer" containerID="8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.391897 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.391935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbxsq\" (UniqueName: \"kubernetes.io/projected/d83e37fd-645e-4258-a441-a9a578844485-kube-api-access-rbxsq\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.391987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-scripts\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.392009 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-config-data\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.392038 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-run-httpd\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.392100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-log-httpd\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.392118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.392136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.395101 4824 scope.go:117] "RemoveContainer" containerID="1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.396302 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-log-httpd\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.396634 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-run-httpd\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.396822 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f\": container with ID starting with 1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f not found: ID does not exist" containerID="1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.396917 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f"} err="failed to get container status \"1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f\": rpc error: code = NotFound desc = could not find container \"1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f\": container with ID starting with 1a12ce8f6f22d1b504493837c6691ee2ea6c4b3fc78eb6341e5e55709d4bad0f not found: ID does not exist" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.397009 4824 scope.go:117] "RemoveContainer" containerID="8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.400977 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-scripts\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.401386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.403159 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933\": container with ID starting with 8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933 not found: ID does not exist" containerID="8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.403193 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933"} err="failed to get container status \"8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933\": rpc error: code = NotFound desc = could not find container \"8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933\": container with ID starting with 8d951a3995513387e2173af97c554426fa5daf201ad803dd0643468987552933 not found: ID does not exist" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.403219 4824 scope.go:117] "RemoveContainer" containerID="8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.407138 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-9sr7j"] Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.409795 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.410317 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-config-data\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.414648 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.415472 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.416777 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbxsq\" (UniqueName: \"kubernetes.io/projected/d83e37fd-645e-4258-a441-a9a578844485-kube-api-access-rbxsq\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.419254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.420230 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-9sr7j"] Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.436584 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.494006 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.494064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-scripts\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.494105 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd9sc\" (UniqueName: \"kubernetes.io/projected/8e361ab6-07b2-49fb-ada0-80eff17ca67b-kube-api-access-cd9sc\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.494208 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-config-data\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.521560 4824 scope.go:117] "RemoveContainer" containerID="742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.573055 4824 scope.go:117] "RemoveContainer" containerID="f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.586277 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.595757 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.595825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-scripts\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.595866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd9sc\" (UniqueName: \"kubernetes.io/projected/8e361ab6-07b2-49fb-ada0-80eff17ca67b-kube-api-access-cd9sc\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.595968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-config-data\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.602561 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-scripts\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.609523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-config-data\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.609764 4824 scope.go:117] "RemoveContainer" containerID="08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.612599 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.657588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd9sc\" (UniqueName: \"kubernetes.io/projected/8e361ab6-07b2-49fb-ada0-80eff17ca67b-kube-api-access-cd9sc\") pod \"nova-cell0-cell-mapping-9sr7j\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.765642 4824 scope.go:117] "RemoveContainer" containerID="8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.780244 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090\": container with ID starting with 8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090 not found: ID does not exist" containerID="8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.780289 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090"} err="failed to get container status \"8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090\": rpc error: code = NotFound desc = could not find container \"8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090\": container with ID starting with 8beb434107b619316c80bd514779fd89c56dcd84c86efebb3120c03867f8e090 not found: ID does not exist" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.780312 4824 scope.go:117] "RemoveContainer" containerID="742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.790536 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2\": container with ID starting with 742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2 not found: ID does not exist" containerID="742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.790576 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2"} err="failed to get container status \"742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2\": rpc error: code = NotFound desc = could not find container \"742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2\": container with ID starting with 742c13256b3a01098379e7b7eb6bdb3d7c9b3656af70e5f1dd2b38640fa339b2 not found: ID does not exist" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.790605 4824 scope.go:117] "RemoveContainer" containerID="f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.800964 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b\": container with ID starting with f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b not found: ID does not exist" containerID="f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.801013 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b"} err="failed to get container status \"f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b\": rpc error: code = NotFound desc = could not find container \"f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b\": container with ID starting with f03e8fe46b0d31e184f5bc669de0c998b15cd14549652d27d568a85fcb41246b not found: ID does not exist" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.801038 4824 scope.go:117] "RemoveContainer" containerID="08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626" Nov 24 13:36:25 crc kubenswrapper[4824]: E1124 13:36:25.813954 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626\": container with ID starting with 08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626 not found: ID does not exist" containerID="08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.814003 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626"} err="failed to get container status \"08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626\": rpc error: code = NotFound desc = could not find container \"08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626\": container with ID starting with 08fbc04b1e39e2182c95ad395025b0e8c0a34e2ee102761505780334c428a626 not found: ID does not exist" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.856292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:25 crc kubenswrapper[4824]: I1124 13:36:25.889442 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:25.998517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.015001 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.059498 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.161788 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.161905 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.166096 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.222776 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.244469 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.246990 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.252901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-config-data\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.253031 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1267dac-9160-4aa1-b487-87f9ad0911f2-logs\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.253084 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.253102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4wk2\" (UniqueName: \"kubernetes.io/projected/a1267dac-9160-4aa1-b487-87f9ad0911f2-kube-api-access-t4wk2\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.259501 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.283665 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.354347 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.355523 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357146 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1267dac-9160-4aa1-b487-87f9ad0911f2-logs\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357179 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gml5\" (UniqueName: \"kubernetes.io/projected/f7779f1e-fae6-493b-81d0-be9a57d310f9-kube-api-access-6gml5\") pod \"nova-scheduler-0\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357205 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4wk2\" (UniqueName: \"kubernetes.io/projected/a1267dac-9160-4aa1-b487-87f9ad0911f2-kube-api-access-t4wk2\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357296 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-logs\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357337 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-config-data\") pod \"nova-scheduler-0\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357356 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-config-data\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357394 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpsxw\" (UniqueName: \"kubernetes.io/projected/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-kube-api-access-fpsxw\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.357421 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-config-data\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.358116 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1267dac-9160-4aa1-b487-87f9ad0911f2-logs\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.360145 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.364875 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.386346 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kgmm4"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.387864 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.391883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-config-data\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.415096 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4wk2\" (UniqueName: \"kubernetes.io/projected/a1267dac-9160-4aa1-b487-87f9ad0911f2-kube-api-access-t4wk2\") pod \"nova-api-0\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.435294 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.443190 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.459854 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gml5\" (UniqueName: \"kubernetes.io/projected/f7779f1e-fae6-493b-81d0-be9a57d310f9-kube-api-access-6gml5\") pod \"nova-scheduler-0\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.459954 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.459990 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-logs\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.460017 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.460034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-config-data\") pod \"nova-scheduler-0\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.460052 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-config-data\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.460074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.460093 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpsxw\" (UniqueName: \"kubernetes.io/projected/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-kube-api-access-fpsxw\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.460111 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.460140 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c67t6\" (UniqueName: \"kubernetes.io/projected/04df1809-4543-4ca9-984f-d3547cbff9d4-kube-api-access-c67t6\") pod \"nova-cell1-novncproxy-0\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.460717 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-logs\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.463009 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kgmm4"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.472405 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.473445 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.477008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-config-data\") pod \"nova-scheduler-0\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.495545 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.514408 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-config-data\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.514949 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gml5\" (UniqueName: \"kubernetes.io/projected/f7779f1e-fae6-493b-81d0-be9a57d310f9-kube-api-access-6gml5\") pod \"nova-scheduler-0\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.517063 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.536196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpsxw\" (UniqueName: \"kubernetes.io/projected/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-kube-api-access-fpsxw\") pod \"nova-metadata-0\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.565860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.565929 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nv78\" (UniqueName: \"kubernetes.io/projected/1a630e6e-19e4-424c-a393-2f08c2c5a154-kube-api-access-8nv78\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.565952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.565986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.566044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.566077 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-config\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.566092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c67t6\" (UniqueName: \"kubernetes.io/projected/04df1809-4543-4ca9-984f-d3547cbff9d4-kube-api-access-c67t6\") pod \"nova-cell1-novncproxy-0\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.566111 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.566155 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-svc\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.581663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.588452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.591858 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.596590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c67t6\" (UniqueName: \"kubernetes.io/projected/04df1809-4543-4ca9-984f-d3547cbff9d4-kube-api-access-c67t6\") pod \"nova-cell1-novncproxy-0\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.661399 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.673076 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-svc\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.674201 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-svc\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.674549 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.675449 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.675611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nv78\" (UniqueName: \"kubernetes.io/projected/1a630e6e-19e4-424c-a393-2f08c2c5a154-kube-api-access-8nv78\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.676081 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.676792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.677058 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-config\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.677715 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-config\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.677769 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.678429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.696454 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nv78\" (UniqueName: \"kubernetes.io/projected/1a630e6e-19e4-424c-a393-2f08c2c5a154-kube-api-access-8nv78\") pod \"dnsmasq-dns-757b4f8459-kgmm4\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.813466 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.879540 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-9sr7j"] Nov 24 13:36:26 crc kubenswrapper[4824]: I1124 13:36:26.975025 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.040917 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28e1651b-29a2-498f-a17e-7d8e50624132" path="/var/lib/kubelet/pods/28e1651b-29a2-498f-a17e-7d8e50624132/volumes" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.041828 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc2d2fae-b656-464e-b1a8-89b4c5160792" path="/var/lib/kubelet/pods/bc2d2fae-b656-464e-b1a8-89b4c5160792/volumes" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.116297 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.146039 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hgk58"] Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.148712 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.152113 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.152260 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.163914 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hgk58"] Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.285268 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.292795 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-config-data\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.302216 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.302417 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-scripts\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.303218 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4294\" (UniqueName: \"kubernetes.io/projected/001160fa-3d5e-4c31-bc99-d8cdd7b81324-kube-api-access-z4294\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.317319 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.355971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61","Type":"ContainerStarted","Data":"2dde950ab7e904b314db9f7ca0bf17fd8d547f3d7bc69b805bec261f0aee8110"} Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.358474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerStarted","Data":"d15f2f855052d5a210caca2e9c6174bcc57899c2489235623a54239e235c008f"} Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.362102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9sr7j" event={"ID":"8e361ab6-07b2-49fb-ada0-80eff17ca67b","Type":"ContainerStarted","Data":"5c5827b6aa081443eea4c5a8b80a7f5900c63746b3dbc07312ffbc7f66fde317"} Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.362153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9sr7j" event={"ID":"8e361ab6-07b2-49fb-ada0-80eff17ca67b","Type":"ContainerStarted","Data":"eb09052250332a0ba5e3f93a23bd42692066bdb9e280cf39f6ae9d89a1f733e2"} Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.365540 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1267dac-9160-4aa1-b487-87f9ad0911f2","Type":"ContainerStarted","Data":"c9202b74865d619fda761ca8df9541b409d2d75abf702fecbccd0695576e1290"} Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.392351 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-9sr7j" podStartSLOduration=2.39233687 podStartE2EDuration="2.39233687s" podCreationTimestamp="2025-11-24 13:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:27.389043489 +0000 UTC m=+1329.028582799" watchObservedRunningTime="2025-11-24 13:36:27.39233687 +0000 UTC m=+1329.031876180" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.404701 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-config-data\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.404785 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.404992 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-scripts\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.405092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4294\" (UniqueName: \"kubernetes.io/projected/001160fa-3d5e-4c31-bc99-d8cdd7b81324-kube-api-access-z4294\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.412295 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-scripts\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.422613 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.428224 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-config-data\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.439287 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4294\" (UniqueName: \"kubernetes.io/projected/001160fa-3d5e-4c31-bc99-d8cdd7b81324-kube-api-access-z4294\") pod \"nova-cell1-conductor-db-sync-hgk58\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.529270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.536124 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:27 crc kubenswrapper[4824]: I1124 13:36:27.795594 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kgmm4"] Nov 24 13:36:28 crc kubenswrapper[4824]: I1124 13:36:28.124489 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hgk58"] Nov 24 13:36:28 crc kubenswrapper[4824]: W1124 13:36:28.150956 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod001160fa_3d5e_4c31_bc99_d8cdd7b81324.slice/crio-bb3ebef6ed3064d847cbdd8d6a024b884ae2e7a0e4648778690801cb1ddf8873 WatchSource:0}: Error finding container bb3ebef6ed3064d847cbdd8d6a024b884ae2e7a0e4648778690801cb1ddf8873: Status 404 returned error can't find the container with id bb3ebef6ed3064d847cbdd8d6a024b884ae2e7a0e4648778690801cb1ddf8873 Nov 24 13:36:28 crc kubenswrapper[4824]: I1124 13:36:28.390405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerStarted","Data":"969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073"} Nov 24 13:36:28 crc kubenswrapper[4824]: I1124 13:36:28.390454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerStarted","Data":"906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049"} Nov 24 13:36:28 crc kubenswrapper[4824]: I1124 13:36:28.395562 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7779f1e-fae6-493b-81d0-be9a57d310f9","Type":"ContainerStarted","Data":"0b5f0e839608b5e11343aa818f003d254eba113d2c08352968722536224b1276"} Nov 24 13:36:28 crc kubenswrapper[4824]: I1124 13:36:28.396588 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"04df1809-4543-4ca9-984f-d3547cbff9d4","Type":"ContainerStarted","Data":"0ac1732af584b7c57910cafa0f1b11d51e7cc8c75ac41fc8e1a0c537b678ecc8"} Nov 24 13:36:28 crc kubenswrapper[4824]: I1124 13:36:28.398086 4824 generic.go:334] "Generic (PLEG): container finished" podID="1a630e6e-19e4-424c-a393-2f08c2c5a154" containerID="545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5" exitCode=0 Nov 24 13:36:28 crc kubenswrapper[4824]: I1124 13:36:28.398134 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" event={"ID":"1a630e6e-19e4-424c-a393-2f08c2c5a154","Type":"ContainerDied","Data":"545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5"} Nov 24 13:36:28 crc kubenswrapper[4824]: I1124 13:36:28.398151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" event={"ID":"1a630e6e-19e4-424c-a393-2f08c2c5a154","Type":"ContainerStarted","Data":"495861b584582646c936fa48b1def5a265eb9e674ba0f9642c4976e713caa9f3"} Nov 24 13:36:28 crc kubenswrapper[4824]: I1124 13:36:28.406064 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-hgk58" event={"ID":"001160fa-3d5e-4c31-bc99-d8cdd7b81324","Type":"ContainerStarted","Data":"bb3ebef6ed3064d847cbdd8d6a024b884ae2e7a0e4648778690801cb1ddf8873"} Nov 24 13:36:29 crc kubenswrapper[4824]: I1124 13:36:29.420677 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerStarted","Data":"d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f"} Nov 24 13:36:29 crc kubenswrapper[4824]: I1124 13:36:29.425499 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" event={"ID":"1a630e6e-19e4-424c-a393-2f08c2c5a154","Type":"ContainerStarted","Data":"8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447"} Nov 24 13:36:29 crc kubenswrapper[4824]: I1124 13:36:29.425964 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:29 crc kubenswrapper[4824]: I1124 13:36:29.438986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-hgk58" event={"ID":"001160fa-3d5e-4c31-bc99-d8cdd7b81324","Type":"ContainerStarted","Data":"8e152f04adaf5d5c412c328288fc0c2b10cf5c58a2c777e18a964a59a7ff59e7"} Nov 24 13:36:29 crc kubenswrapper[4824]: I1124 13:36:29.459619 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" podStartSLOduration=3.459604262 podStartE2EDuration="3.459604262s" podCreationTimestamp="2025-11-24 13:36:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:29.453317615 +0000 UTC m=+1331.092856925" watchObservedRunningTime="2025-11-24 13:36:29.459604262 +0000 UTC m=+1331.099143572" Nov 24 13:36:29 crc kubenswrapper[4824]: I1124 13:36:29.472002 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-hgk58" podStartSLOduration=2.471987332 podStartE2EDuration="2.471987332s" podCreationTimestamp="2025-11-24 13:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:29.469224802 +0000 UTC m=+1331.108764112" watchObservedRunningTime="2025-11-24 13:36:29.471987332 +0000 UTC m=+1331.111526642" Nov 24 13:36:29 crc kubenswrapper[4824]: I1124 13:36:29.805022 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:36:29 crc kubenswrapper[4824]: I1124 13:36:29.819243 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:36:30 crc kubenswrapper[4824]: I1124 13:36:30.596307 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.567997 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61","Type":"ContainerStarted","Data":"543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f"} Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.568603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61","Type":"ContainerStarted","Data":"5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa"} Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.568374 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerName="nova-metadata-metadata" containerID="cri-o://543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f" gracePeriod=30 Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.568065 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerName="nova-metadata-log" containerID="cri-o://5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa" gracePeriod=30 Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.574519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerStarted","Data":"7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea"} Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.574644 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.578214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7779f1e-fae6-493b-81d0-be9a57d310f9","Type":"ContainerStarted","Data":"75c44665ccbe357a4934e151d2bca95543425285f24204ea3fd2054a8d3a2cec"} Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.581383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1267dac-9160-4aa1-b487-87f9ad0911f2","Type":"ContainerStarted","Data":"721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3"} Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.581418 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1267dac-9160-4aa1-b487-87f9ad0911f2","Type":"ContainerStarted","Data":"089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765"} Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.585623 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.4662310720000002 podStartE2EDuration="8.585606947s" podCreationTimestamp="2025-11-24 13:36:25 +0000 UTC" firstStartedPulling="2025-11-24 13:36:27.284674565 +0000 UTC m=+1328.924213875" lastFinishedPulling="2025-11-24 13:36:32.40405044 +0000 UTC m=+1334.043589750" observedRunningTime="2025-11-24 13:36:33.583385639 +0000 UTC m=+1335.222924949" watchObservedRunningTime="2025-11-24 13:36:33.585606947 +0000 UTC m=+1335.225146257" Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.594147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"04df1809-4543-4ca9-984f-d3547cbff9d4","Type":"ContainerStarted","Data":"0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795"} Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.594268 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="04df1809-4543-4ca9-984f-d3547cbff9d4" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795" gracePeriod=30 Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.678762 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.690507975 podStartE2EDuration="8.678742276s" podCreationTimestamp="2025-11-24 13:36:25 +0000 UTC" firstStartedPulling="2025-11-24 13:36:26.44294093 +0000 UTC m=+1328.082480240" lastFinishedPulling="2025-11-24 13:36:32.431175231 +0000 UTC m=+1334.070714541" observedRunningTime="2025-11-24 13:36:33.63254055 +0000 UTC m=+1335.272079860" watchObservedRunningTime="2025-11-24 13:36:33.678742276 +0000 UTC m=+1335.318281586" Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.706473 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.482298092 podStartE2EDuration="8.7064533s" podCreationTimestamp="2025-11-24 13:36:25 +0000 UTC" firstStartedPulling="2025-11-24 13:36:27.200251026 +0000 UTC m=+1328.839790336" lastFinishedPulling="2025-11-24 13:36:32.424406234 +0000 UTC m=+1334.063945544" observedRunningTime="2025-11-24 13:36:33.606464332 +0000 UTC m=+1335.246003642" watchObservedRunningTime="2025-11-24 13:36:33.7064533 +0000 UTC m=+1335.345992610" Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.710370 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.898408017 podStartE2EDuration="8.710352285s" podCreationTimestamp="2025-11-24 13:36:25 +0000 UTC" firstStartedPulling="2025-11-24 13:36:27.613155661 +0000 UTC m=+1329.252694971" lastFinishedPulling="2025-11-24 13:36:32.425099929 +0000 UTC m=+1334.064639239" observedRunningTime="2025-11-24 13:36:33.660295214 +0000 UTC m=+1335.299834524" watchObservedRunningTime="2025-11-24 13:36:33.710352285 +0000 UTC m=+1335.349891595" Nov 24 13:36:33 crc kubenswrapper[4824]: I1124 13:36:33.731767 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.672302621 podStartE2EDuration="7.73171546s" podCreationTimestamp="2025-11-24 13:36:26 +0000 UTC" firstStartedPulling="2025-11-24 13:36:27.355559639 +0000 UTC m=+1328.995098949" lastFinishedPulling="2025-11-24 13:36:32.414972478 +0000 UTC m=+1334.054511788" observedRunningTime="2025-11-24 13:36:33.681930805 +0000 UTC m=+1335.321470115" watchObservedRunningTime="2025-11-24 13:36:33.73171546 +0000 UTC m=+1335.371254770" Nov 24 13:36:34 crc kubenswrapper[4824]: I1124 13:36:34.609194 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerID="5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa" exitCode=143 Nov 24 13:36:34 crc kubenswrapper[4824]: I1124 13:36:34.610192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61","Type":"ContainerDied","Data":"5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa"} Nov 24 13:36:36 crc kubenswrapper[4824]: I1124 13:36:36.473552 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:36:36 crc kubenswrapper[4824]: I1124 13:36:36.473841 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:36:36 crc kubenswrapper[4824]: I1124 13:36:36.592692 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:36:36 crc kubenswrapper[4824]: I1124 13:36:36.592741 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:36:36 crc kubenswrapper[4824]: I1124 13:36:36.663655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:36:36 crc kubenswrapper[4824]: I1124 13:36:36.814788 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4824]: I1124 13:36:36.814876 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4824]: I1124 13:36:36.847492 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 13:36:36 crc kubenswrapper[4824]: I1124 13:36:36.977598 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.034944 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-k264z"] Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.035174 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" podUID="0367b064-fb20-459e-8878-e60a5c8b8207" containerName="dnsmasq-dns" containerID="cri-o://d477037c13bc616f87cccca7d9817b438eb3db455aea1620a5f6b7a903937b9a" gracePeriod=10 Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.575360 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.576132 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.670089 4824 generic.go:334] "Generic (PLEG): container finished" podID="0367b064-fb20-459e-8878-e60a5c8b8207" containerID="d477037c13bc616f87cccca7d9817b438eb3db455aea1620a5f6b7a903937b9a" exitCode=0 Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.670198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" event={"ID":"0367b064-fb20-459e-8878-e60a5c8b8207","Type":"ContainerDied","Data":"d477037c13bc616f87cccca7d9817b438eb3db455aea1620a5f6b7a903937b9a"} Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.728153 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.851989 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.947515 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rhbl\" (UniqueName: \"kubernetes.io/projected/0367b064-fb20-459e-8878-e60a5c8b8207-kube-api-access-2rhbl\") pod \"0367b064-fb20-459e-8878-e60a5c8b8207\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.947658 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-svc\") pod \"0367b064-fb20-459e-8878-e60a5c8b8207\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.947693 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-config\") pod \"0367b064-fb20-459e-8878-e60a5c8b8207\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.947796 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-sb\") pod \"0367b064-fb20-459e-8878-e60a5c8b8207\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.947833 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-nb\") pod \"0367b064-fb20-459e-8878-e60a5c8b8207\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.947917 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-swift-storage-0\") pod \"0367b064-fb20-459e-8878-e60a5c8b8207\" (UID: \"0367b064-fb20-459e-8878-e60a5c8b8207\") " Nov 24 13:36:37 crc kubenswrapper[4824]: I1124 13:36:37.974941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0367b064-fb20-459e-8878-e60a5c8b8207-kube-api-access-2rhbl" (OuterVolumeSpecName: "kube-api-access-2rhbl") pod "0367b064-fb20-459e-8878-e60a5c8b8207" (UID: "0367b064-fb20-459e-8878-e60a5c8b8207"). InnerVolumeSpecName "kube-api-access-2rhbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.052028 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rhbl\" (UniqueName: \"kubernetes.io/projected/0367b064-fb20-459e-8878-e60a5c8b8207-kube-api-access-2rhbl\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.060374 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0367b064-fb20-459e-8878-e60a5c8b8207" (UID: "0367b064-fb20-459e-8878-e60a5c8b8207"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.068059 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-config" (OuterVolumeSpecName: "config") pod "0367b064-fb20-459e-8878-e60a5c8b8207" (UID: "0367b064-fb20-459e-8878-e60a5c8b8207"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.085962 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0367b064-fb20-459e-8878-e60a5c8b8207" (UID: "0367b064-fb20-459e-8878-e60a5c8b8207"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.106426 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0367b064-fb20-459e-8878-e60a5c8b8207" (UID: "0367b064-fb20-459e-8878-e60a5c8b8207"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.110773 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0367b064-fb20-459e-8878-e60a5c8b8207" (UID: "0367b064-fb20-459e-8878-e60a5c8b8207"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.154736 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.155001 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.155075 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.155136 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.155391 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0367b064-fb20-459e-8878-e60a5c8b8207-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.679282 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.681870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-k264z" event={"ID":"0367b064-fb20-459e-8878-e60a5c8b8207","Type":"ContainerDied","Data":"3b28efe5d0535be381ef997adc0445b7bd1217bb881542e1897516de79c576e2"} Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.681938 4824 scope.go:117] "RemoveContainer" containerID="d477037c13bc616f87cccca7d9817b438eb3db455aea1620a5f6b7a903937b9a" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.738864 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-k264z"] Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.748281 4824 scope.go:117] "RemoveContainer" containerID="6bcad28ae4ef545e81c03a052f5b1a3502d9d111cf5add261503dae1bb02340c" Nov 24 13:36:38 crc kubenswrapper[4824]: I1124 13:36:38.751738 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-k264z"] Nov 24 13:36:39 crc kubenswrapper[4824]: I1124 13:36:39.021009 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0367b064-fb20-459e-8878-e60a5c8b8207" path="/var/lib/kubelet/pods/0367b064-fb20-459e-8878-e60a5c8b8207/volumes" Nov 24 13:36:39 crc kubenswrapper[4824]: I1124 13:36:39.695664 4824 generic.go:334] "Generic (PLEG): container finished" podID="8e361ab6-07b2-49fb-ada0-80eff17ca67b" containerID="5c5827b6aa081443eea4c5a8b80a7f5900c63746b3dbc07312ffbc7f66fde317" exitCode=0 Nov 24 13:36:39 crc kubenswrapper[4824]: I1124 13:36:39.695702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9sr7j" event={"ID":"8e361ab6-07b2-49fb-ada0-80eff17ca67b","Type":"ContainerDied","Data":"5c5827b6aa081443eea4c5a8b80a7f5900c63746b3dbc07312ffbc7f66fde317"} Nov 24 13:36:40 crc kubenswrapper[4824]: I1124 13:36:40.706304 4824 generic.go:334] "Generic (PLEG): container finished" podID="001160fa-3d5e-4c31-bc99-d8cdd7b81324" containerID="8e152f04adaf5d5c412c328288fc0c2b10cf5c58a2c777e18a964a59a7ff59e7" exitCode=0 Nov 24 13:36:40 crc kubenswrapper[4824]: I1124 13:36:40.706997 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-hgk58" event={"ID":"001160fa-3d5e-4c31-bc99-d8cdd7b81324","Type":"ContainerDied","Data":"8e152f04adaf5d5c412c328288fc0c2b10cf5c58a2c777e18a964a59a7ff59e7"} Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.100547 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.149387 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-config-data\") pod \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.149432 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-combined-ca-bundle\") pod \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.149524 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-scripts\") pod \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.149556 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd9sc\" (UniqueName: \"kubernetes.io/projected/8e361ab6-07b2-49fb-ada0-80eff17ca67b-kube-api-access-cd9sc\") pod \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\" (UID: \"8e361ab6-07b2-49fb-ada0-80eff17ca67b\") " Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.182253 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-scripts" (OuterVolumeSpecName: "scripts") pod "8e361ab6-07b2-49fb-ada0-80eff17ca67b" (UID: "8e361ab6-07b2-49fb-ada0-80eff17ca67b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.183956 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e361ab6-07b2-49fb-ada0-80eff17ca67b-kube-api-access-cd9sc" (OuterVolumeSpecName: "kube-api-access-cd9sc") pod "8e361ab6-07b2-49fb-ada0-80eff17ca67b" (UID: "8e361ab6-07b2-49fb-ada0-80eff17ca67b"). InnerVolumeSpecName "kube-api-access-cd9sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.218817 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-config-data" (OuterVolumeSpecName: "config-data") pod "8e361ab6-07b2-49fb-ada0-80eff17ca67b" (UID: "8e361ab6-07b2-49fb-ada0-80eff17ca67b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.231088 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e361ab6-07b2-49fb-ada0-80eff17ca67b" (UID: "8e361ab6-07b2-49fb-ada0-80eff17ca67b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.253133 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.253850 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.253968 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e361ab6-07b2-49fb-ada0-80eff17ca67b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.254065 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd9sc\" (UniqueName: \"kubernetes.io/projected/8e361ab6-07b2-49fb-ada0-80eff17ca67b-kube-api-access-cd9sc\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.723524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9sr7j" event={"ID":"8e361ab6-07b2-49fb-ada0-80eff17ca67b","Type":"ContainerDied","Data":"eb09052250332a0ba5e3f93a23bd42692066bdb9e280cf39f6ae9d89a1f733e2"} Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.724661 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb09052250332a0ba5e3f93a23bd42692066bdb9e280cf39f6ae9d89a1f733e2" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.723643 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9sr7j" Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.917936 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.918186 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-log" containerID="cri-o://089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765" gracePeriod=30 Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.918553 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-api" containerID="cri-o://721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3" gracePeriod=30 Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.942524 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:41 crc kubenswrapper[4824]: I1124 13:36:41.942734 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f7779f1e-fae6-493b-81d0-be9a57d310f9" containerName="nova-scheduler-scheduler" containerID="cri-o://75c44665ccbe357a4934e151d2bca95543425285f24204ea3fd2054a8d3a2cec" gracePeriod=30 Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.111671 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.172311 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-config-data\") pod \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.172450 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-scripts\") pod \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.172494 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-combined-ca-bundle\") pod \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.172565 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4294\" (UniqueName: \"kubernetes.io/projected/001160fa-3d5e-4c31-bc99-d8cdd7b81324-kube-api-access-z4294\") pod \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\" (UID: \"001160fa-3d5e-4c31-bc99-d8cdd7b81324\") " Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.179057 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001160fa-3d5e-4c31-bc99-d8cdd7b81324-kube-api-access-z4294" (OuterVolumeSpecName: "kube-api-access-z4294") pod "001160fa-3d5e-4c31-bc99-d8cdd7b81324" (UID: "001160fa-3d5e-4c31-bc99-d8cdd7b81324"). InnerVolumeSpecName "kube-api-access-z4294". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.193338 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-scripts" (OuterVolumeSpecName: "scripts") pod "001160fa-3d5e-4c31-bc99-d8cdd7b81324" (UID: "001160fa-3d5e-4c31-bc99-d8cdd7b81324"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.221566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "001160fa-3d5e-4c31-bc99-d8cdd7b81324" (UID: "001160fa-3d5e-4c31-bc99-d8cdd7b81324"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.222429 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-config-data" (OuterVolumeSpecName: "config-data") pod "001160fa-3d5e-4c31-bc99-d8cdd7b81324" (UID: "001160fa-3d5e-4c31-bc99-d8cdd7b81324"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.274841 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.274877 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.274886 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001160fa-3d5e-4c31-bc99-d8cdd7b81324-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.274897 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4294\" (UniqueName: \"kubernetes.io/projected/001160fa-3d5e-4c31-bc99-d8cdd7b81324-kube-api-access-z4294\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.732835 4824 generic.go:334] "Generic (PLEG): container finished" podID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerID="089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765" exitCode=143 Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.732920 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1267dac-9160-4aa1-b487-87f9ad0911f2","Type":"ContainerDied","Data":"089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765"} Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.734744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-hgk58" event={"ID":"001160fa-3d5e-4c31-bc99-d8cdd7b81324","Type":"ContainerDied","Data":"bb3ebef6ed3064d847cbdd8d6a024b884ae2e7a0e4648778690801cb1ddf8873"} Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.734782 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb3ebef6ed3064d847cbdd8d6a024b884ae2e7a0e4648778690801cb1ddf8873" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.734843 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-hgk58" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.835969 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 13:36:42 crc kubenswrapper[4824]: E1124 13:36:42.836346 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0367b064-fb20-459e-8878-e60a5c8b8207" containerName="init" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.836361 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0367b064-fb20-459e-8878-e60a5c8b8207" containerName="init" Nov 24 13:36:42 crc kubenswrapper[4824]: E1124 13:36:42.836368 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0367b064-fb20-459e-8878-e60a5c8b8207" containerName="dnsmasq-dns" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.836374 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0367b064-fb20-459e-8878-e60a5c8b8207" containerName="dnsmasq-dns" Nov 24 13:36:42 crc kubenswrapper[4824]: E1124 13:36:42.836391 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001160fa-3d5e-4c31-bc99-d8cdd7b81324" containerName="nova-cell1-conductor-db-sync" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.836397 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="001160fa-3d5e-4c31-bc99-d8cdd7b81324" containerName="nova-cell1-conductor-db-sync" Nov 24 13:36:42 crc kubenswrapper[4824]: E1124 13:36:42.836412 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e361ab6-07b2-49fb-ada0-80eff17ca67b" containerName="nova-manage" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.836417 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e361ab6-07b2-49fb-ada0-80eff17ca67b" containerName="nova-manage" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.836584 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0367b064-fb20-459e-8878-e60a5c8b8207" containerName="dnsmasq-dns" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.836602 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e361ab6-07b2-49fb-ada0-80eff17ca67b" containerName="nova-manage" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.836610 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="001160fa-3d5e-4c31-bc99-d8cdd7b81324" containerName="nova-cell1-conductor-db-sync" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.837227 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.840758 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.854443 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.987569 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73aeee43-4360-4af8-9c1e-e1d9bc712157-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"73aeee43-4360-4af8-9c1e-e1d9bc712157\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.988067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h6hf\" (UniqueName: \"kubernetes.io/projected/73aeee43-4360-4af8-9c1e-e1d9bc712157-kube-api-access-7h6hf\") pod \"nova-cell1-conductor-0\" (UID: \"73aeee43-4360-4af8-9c1e-e1d9bc712157\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:42 crc kubenswrapper[4824]: I1124 13:36:42.988244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73aeee43-4360-4af8-9c1e-e1d9bc712157-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"73aeee43-4360-4af8-9c1e-e1d9bc712157\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:43 crc kubenswrapper[4824]: I1124 13:36:43.090079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h6hf\" (UniqueName: \"kubernetes.io/projected/73aeee43-4360-4af8-9c1e-e1d9bc712157-kube-api-access-7h6hf\") pod \"nova-cell1-conductor-0\" (UID: \"73aeee43-4360-4af8-9c1e-e1d9bc712157\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:43 crc kubenswrapper[4824]: I1124 13:36:43.090241 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73aeee43-4360-4af8-9c1e-e1d9bc712157-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"73aeee43-4360-4af8-9c1e-e1d9bc712157\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:43 crc kubenswrapper[4824]: I1124 13:36:43.090297 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73aeee43-4360-4af8-9c1e-e1d9bc712157-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"73aeee43-4360-4af8-9c1e-e1d9bc712157\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:43 crc kubenswrapper[4824]: I1124 13:36:43.095440 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73aeee43-4360-4af8-9c1e-e1d9bc712157-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"73aeee43-4360-4af8-9c1e-e1d9bc712157\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:43 crc kubenswrapper[4824]: I1124 13:36:43.097409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73aeee43-4360-4af8-9c1e-e1d9bc712157-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"73aeee43-4360-4af8-9c1e-e1d9bc712157\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:43 crc kubenswrapper[4824]: I1124 13:36:43.118499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h6hf\" (UniqueName: \"kubernetes.io/projected/73aeee43-4360-4af8-9c1e-e1d9bc712157-kube-api-access-7h6hf\") pod \"nova-cell1-conductor-0\" (UID: \"73aeee43-4360-4af8-9c1e-e1d9bc712157\") " pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:43 crc kubenswrapper[4824]: I1124 13:36:43.151717 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:43 crc kubenswrapper[4824]: I1124 13:36:43.611619 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 13:36:43 crc kubenswrapper[4824]: W1124 13:36:43.623572 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73aeee43_4360_4af8_9c1e_e1d9bc712157.slice/crio-5bee76c29c96e4e615e76e5f1c785fe7477890e28822617521ef568fb55a0eec WatchSource:0}: Error finding container 5bee76c29c96e4e615e76e5f1c785fe7477890e28822617521ef568fb55a0eec: Status 404 returned error can't find the container with id 5bee76c29c96e4e615e76e5f1c785fe7477890e28822617521ef568fb55a0eec Nov 24 13:36:43 crc kubenswrapper[4824]: I1124 13:36:43.745065 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"73aeee43-4360-4af8-9c1e-e1d9bc712157","Type":"ContainerStarted","Data":"5bee76c29c96e4e615e76e5f1c785fe7477890e28822617521ef568fb55a0eec"} Nov 24 13:36:44 crc kubenswrapper[4824]: I1124 13:36:44.756280 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"73aeee43-4360-4af8-9c1e-e1d9bc712157","Type":"ContainerStarted","Data":"ffc8666479722c105962c2887e725b4db9a7d9b5a1cbce8b2d3ddcd911a28a69"} Nov 24 13:36:44 crc kubenswrapper[4824]: I1124 13:36:44.758198 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:44 crc kubenswrapper[4824]: I1124 13:36:44.779738 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.7797198659999998 podStartE2EDuration="2.779719866s" podCreationTimestamp="2025-11-24 13:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:44.77712811 +0000 UTC m=+1346.416667460" watchObservedRunningTime="2025-11-24 13:36:44.779719866 +0000 UTC m=+1346.419259186" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.526075 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.638257 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-config-data\") pod \"a1267dac-9160-4aa1-b487-87f9ad0911f2\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.638419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-combined-ca-bundle\") pod \"a1267dac-9160-4aa1-b487-87f9ad0911f2\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.638571 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4wk2\" (UniqueName: \"kubernetes.io/projected/a1267dac-9160-4aa1-b487-87f9ad0911f2-kube-api-access-t4wk2\") pod \"a1267dac-9160-4aa1-b487-87f9ad0911f2\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.638610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1267dac-9160-4aa1-b487-87f9ad0911f2-logs\") pod \"a1267dac-9160-4aa1-b487-87f9ad0911f2\" (UID: \"a1267dac-9160-4aa1-b487-87f9ad0911f2\") " Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.639449 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1267dac-9160-4aa1-b487-87f9ad0911f2-logs" (OuterVolumeSpecName: "logs") pod "a1267dac-9160-4aa1-b487-87f9ad0911f2" (UID: "a1267dac-9160-4aa1-b487-87f9ad0911f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.645919 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1267dac-9160-4aa1-b487-87f9ad0911f2-kube-api-access-t4wk2" (OuterVolumeSpecName: "kube-api-access-t4wk2") pod "a1267dac-9160-4aa1-b487-87f9ad0911f2" (UID: "a1267dac-9160-4aa1-b487-87f9ad0911f2"). InnerVolumeSpecName "kube-api-access-t4wk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.712421 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-config-data" (OuterVolumeSpecName: "config-data") pod "a1267dac-9160-4aa1-b487-87f9ad0911f2" (UID: "a1267dac-9160-4aa1-b487-87f9ad0911f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.716847 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1267dac-9160-4aa1-b487-87f9ad0911f2" (UID: "a1267dac-9160-4aa1-b487-87f9ad0911f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.741364 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4wk2\" (UniqueName: \"kubernetes.io/projected/a1267dac-9160-4aa1-b487-87f9ad0911f2-kube-api-access-t4wk2\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.741401 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1267dac-9160-4aa1-b487-87f9ad0911f2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.741415 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.741429 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1267dac-9160-4aa1-b487-87f9ad0911f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.766317 4824 generic.go:334] "Generic (PLEG): container finished" podID="f7779f1e-fae6-493b-81d0-be9a57d310f9" containerID="75c44665ccbe357a4934e151d2bca95543425285f24204ea3fd2054a8d3a2cec" exitCode=0 Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.766577 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7779f1e-fae6-493b-81d0-be9a57d310f9","Type":"ContainerDied","Data":"75c44665ccbe357a4934e151d2bca95543425285f24204ea3fd2054a8d3a2cec"} Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.766624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7779f1e-fae6-493b-81d0-be9a57d310f9","Type":"ContainerDied","Data":"0b5f0e839608b5e11343aa818f003d254eba113d2c08352968722536224b1276"} Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.766639 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b5f0e839608b5e11343aa818f003d254eba113d2c08352968722536224b1276" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.768297 4824 generic.go:334] "Generic (PLEG): container finished" podID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerID="721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3" exitCode=0 Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.768568 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.769208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1267dac-9160-4aa1-b487-87f9ad0911f2","Type":"ContainerDied","Data":"721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3"} Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.769239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1267dac-9160-4aa1-b487-87f9ad0911f2","Type":"ContainerDied","Data":"c9202b74865d619fda761ca8df9541b409d2d75abf702fecbccd0695576e1290"} Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.769259 4824 scope.go:117] "RemoveContainer" containerID="721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.786343 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.833351 4824 scope.go:117] "RemoveContainer" containerID="089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.842507 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gml5\" (UniqueName: \"kubernetes.io/projected/f7779f1e-fae6-493b-81d0-be9a57d310f9-kube-api-access-6gml5\") pod \"f7779f1e-fae6-493b-81d0-be9a57d310f9\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.842559 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-config-data\") pod \"f7779f1e-fae6-493b-81d0-be9a57d310f9\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.842643 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-combined-ca-bundle\") pod \"f7779f1e-fae6-493b-81d0-be9a57d310f9\" (UID: \"f7779f1e-fae6-493b-81d0-be9a57d310f9\") " Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.862107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7779f1e-fae6-493b-81d0-be9a57d310f9-kube-api-access-6gml5" (OuterVolumeSpecName: "kube-api-access-6gml5") pod "f7779f1e-fae6-493b-81d0-be9a57d310f9" (UID: "f7779f1e-fae6-493b-81d0-be9a57d310f9"). InnerVolumeSpecName "kube-api-access-6gml5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.869945 4824 scope.go:117] "RemoveContainer" containerID="721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3" Nov 24 13:36:45 crc kubenswrapper[4824]: E1124 13:36:45.871747 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3\": container with ID starting with 721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3 not found: ID does not exist" containerID="721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.871777 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3"} err="failed to get container status \"721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3\": rpc error: code = NotFound desc = could not find container \"721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3\": container with ID starting with 721c23e0d18904762dd77a6280f323c409b28d16b0aaeb4fb8105cfc11b73fc3 not found: ID does not exist" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.871795 4824 scope.go:117] "RemoveContainer" containerID="089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.871862 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:45 crc kubenswrapper[4824]: E1124 13:36:45.873564 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765\": container with ID starting with 089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765 not found: ID does not exist" containerID="089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.873590 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765"} err="failed to get container status \"089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765\": rpc error: code = NotFound desc = could not find container \"089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765\": container with ID starting with 089f7627d3083d6718e8ff909ab5bc0454db1aba858d83d42f7eac300f513765 not found: ID does not exist" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.881564 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.888955 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:45 crc kubenswrapper[4824]: E1124 13:36:45.889358 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7779f1e-fae6-493b-81d0-be9a57d310f9" containerName="nova-scheduler-scheduler" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.889373 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7779f1e-fae6-493b-81d0-be9a57d310f9" containerName="nova-scheduler-scheduler" Nov 24 13:36:45 crc kubenswrapper[4824]: E1124 13:36:45.889412 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-log" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.889419 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-log" Nov 24 13:36:45 crc kubenswrapper[4824]: E1124 13:36:45.889430 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-api" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.889436 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-api" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.889597 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-api" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.889613 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" containerName="nova-api-log" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.889620 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7779f1e-fae6-493b-81d0-be9a57d310f9" containerName="nova-scheduler-scheduler" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.890574 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.892788 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.897083 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-config-data" (OuterVolumeSpecName: "config-data") pod "f7779f1e-fae6-493b-81d0-be9a57d310f9" (UID: "f7779f1e-fae6-493b-81d0-be9a57d310f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.898832 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.922310 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7779f1e-fae6-493b-81d0-be9a57d310f9" (UID: "f7779f1e-fae6-493b-81d0-be9a57d310f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.951993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f2c2432-5906-40b9-b6ff-795783620bc2-logs\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.952076 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.952172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-config-data\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.952236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrbvk\" (UniqueName: \"kubernetes.io/projected/9f2c2432-5906-40b9-b6ff-795783620bc2-kube-api-access-xrbvk\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.952519 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.952548 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gml5\" (UniqueName: \"kubernetes.io/projected/f7779f1e-fae6-493b-81d0-be9a57d310f9-kube-api-access-6gml5\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:45 crc kubenswrapper[4824]: I1124 13:36:45.952564 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7779f1e-fae6-493b-81d0-be9a57d310f9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.054234 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.054308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-config-data\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.054356 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrbvk\" (UniqueName: \"kubernetes.io/projected/9f2c2432-5906-40b9-b6ff-795783620bc2-kube-api-access-xrbvk\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.054431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f2c2432-5906-40b9-b6ff-795783620bc2-logs\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.055067 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f2c2432-5906-40b9-b6ff-795783620bc2-logs\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.058000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.058461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-config-data\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.070895 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrbvk\" (UniqueName: \"kubernetes.io/projected/9f2c2432-5906-40b9-b6ff-795783620bc2-kube-api-access-xrbvk\") pod \"nova-api-0\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " pod="openstack/nova-api-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.223309 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.710466 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:36:46 crc kubenswrapper[4824]: W1124 13:36:46.716599 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f2c2432_5906_40b9_b6ff_795783620bc2.slice/crio-5273f19b62974ab23ac94ddd12bb7cebea98e515cec50a990f5641d8656acbd0 WatchSource:0}: Error finding container 5273f19b62974ab23ac94ddd12bb7cebea98e515cec50a990f5641d8656acbd0: Status 404 returned error can't find the container with id 5273f19b62974ab23ac94ddd12bb7cebea98e515cec50a990f5641d8656acbd0 Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.783623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f2c2432-5906-40b9-b6ff-795783620bc2","Type":"ContainerStarted","Data":"5273f19b62974ab23ac94ddd12bb7cebea98e515cec50a990f5641d8656acbd0"} Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.783709 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.869030 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.883863 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.895319 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.896716 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.903132 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.916416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.971720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.971985 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nkkd\" (UniqueName: \"kubernetes.io/projected/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-kube-api-access-4nkkd\") pod \"nova-scheduler-0\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:46 crc kubenswrapper[4824]: I1124 13:36:46.972035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-config-data\") pod \"nova-scheduler-0\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.047628 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1267dac-9160-4aa1-b487-87f9ad0911f2" path="/var/lib/kubelet/pods/a1267dac-9160-4aa1-b487-87f9ad0911f2/volumes" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.048422 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7779f1e-fae6-493b-81d0-be9a57d310f9" path="/var/lib/kubelet/pods/f7779f1e-fae6-493b-81d0-be9a57d310f9/volumes" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.073230 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nkkd\" (UniqueName: \"kubernetes.io/projected/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-kube-api-access-4nkkd\") pod \"nova-scheduler-0\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.073296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-config-data\") pod \"nova-scheduler-0\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.073419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.081864 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.081744 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-config-data\") pod \"nova-scheduler-0\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.094989 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nkkd\" (UniqueName: \"kubernetes.io/projected/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-kube-api-access-4nkkd\") pod \"nova-scheduler-0\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " pod="openstack/nova-scheduler-0" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.228365 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.667647 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:36:47 crc kubenswrapper[4824]: W1124 13:36:47.670446 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9661f7ee_75ed_4c09_be2d_d08d2c8eda1b.slice/crio-a68568fa73f85e68753822e99baa3d533c77a0dc0075d98a227a5885cda1c246 WatchSource:0}: Error finding container a68568fa73f85e68753822e99baa3d533c77a0dc0075d98a227a5885cda1c246: Status 404 returned error can't find the container with id a68568fa73f85e68753822e99baa3d533c77a0dc0075d98a227a5885cda1c246 Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.800225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f2c2432-5906-40b9-b6ff-795783620bc2","Type":"ContainerStarted","Data":"edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd"} Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.800474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f2c2432-5906-40b9-b6ff-795783620bc2","Type":"ContainerStarted","Data":"2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac"} Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.803621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b","Type":"ContainerStarted","Data":"a68568fa73f85e68753822e99baa3d533c77a0dc0075d98a227a5885cda1c246"} Nov 24 13:36:47 crc kubenswrapper[4824]: I1124 13:36:47.823412 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.823396895 podStartE2EDuration="2.823396895s" podCreationTimestamp="2025-11-24 13:36:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:47.813576511 +0000 UTC m=+1349.453115831" watchObservedRunningTime="2025-11-24 13:36:47.823396895 +0000 UTC m=+1349.462936205" Nov 24 13:36:48 crc kubenswrapper[4824]: I1124 13:36:48.179287 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 13:36:49 crc kubenswrapper[4824]: I1124 13:36:49.091857 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b","Type":"ContainerStarted","Data":"fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39"} Nov 24 13:36:49 crc kubenswrapper[4824]: I1124 13:36:49.118384 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.118368463 podStartE2EDuration="3.118368463s" podCreationTimestamp="2025-11-24 13:36:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:36:49.115125673 +0000 UTC m=+1350.754665023" watchObservedRunningTime="2025-11-24 13:36:49.118368463 +0000 UTC m=+1350.757907763" Nov 24 13:36:52 crc kubenswrapper[4824]: I1124 13:36:52.228978 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 13:36:55 crc kubenswrapper[4824]: I1124 13:36:55.600322 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 13:36:56 crc kubenswrapper[4824]: I1124 13:36:56.223869 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:36:56 crc kubenswrapper[4824]: I1124 13:36:56.223936 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:36:57 crc kubenswrapper[4824]: I1124 13:36:57.229469 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 13:36:57 crc kubenswrapper[4824]: I1124 13:36:57.277444 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 13:36:57 crc kubenswrapper[4824]: I1124 13:36:57.307982 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:57 crc kubenswrapper[4824]: I1124 13:36:57.308024 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 13:36:58 crc kubenswrapper[4824]: I1124 13:36:58.226608 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.106771 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.112436 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.244519 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpsxw\" (UniqueName: \"kubernetes.io/projected/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-kube-api-access-fpsxw\") pod \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.245469 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c67t6\" (UniqueName: \"kubernetes.io/projected/04df1809-4543-4ca9-984f-d3547cbff9d4-kube-api-access-c67t6\") pod \"04df1809-4543-4ca9-984f-d3547cbff9d4\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.245745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-combined-ca-bundle\") pod \"04df1809-4543-4ca9-984f-d3547cbff9d4\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.245871 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-logs\") pod \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.245988 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-config-data\") pod \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.246194 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-combined-ca-bundle\") pod \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\" (UID: \"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61\") " Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.246610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-config-data\") pod \"04df1809-4543-4ca9-984f-d3547cbff9d4\" (UID: \"04df1809-4543-4ca9-984f-d3547cbff9d4\") " Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.246476 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-logs" (OuterVolumeSpecName: "logs") pod "f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" (UID: "f4f69ff2-19d7-4a2b-a34e-75062ed8ff61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.251528 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04df1809-4543-4ca9-984f-d3547cbff9d4-kube-api-access-c67t6" (OuterVolumeSpecName: "kube-api-access-c67t6") pod "04df1809-4543-4ca9-984f-d3547cbff9d4" (UID: "04df1809-4543-4ca9-984f-d3547cbff9d4"). InnerVolumeSpecName "kube-api-access-c67t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.254960 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-kube-api-access-fpsxw" (OuterVolumeSpecName: "kube-api-access-fpsxw") pod "f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" (UID: "f4f69ff2-19d7-4a2b-a34e-75062ed8ff61"). InnerVolumeSpecName "kube-api-access-fpsxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.259771 4824 generic.go:334] "Generic (PLEG): container finished" podID="04df1809-4543-4ca9-984f-d3547cbff9d4" containerID="0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795" exitCode=137 Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.259842 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"04df1809-4543-4ca9-984f-d3547cbff9d4","Type":"ContainerDied","Data":"0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795"} Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.259868 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"04df1809-4543-4ca9-984f-d3547cbff9d4","Type":"ContainerDied","Data":"0ac1732af584b7c57910cafa0f1b11d51e7cc8c75ac41fc8e1a0c537b678ecc8"} Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.259884 4824 scope.go:117] "RemoveContainer" containerID="0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.259901 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.265118 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerID="543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f" exitCode=137 Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.265168 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61","Type":"ContainerDied","Data":"543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f"} Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.265194 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4f69ff2-19d7-4a2b-a34e-75062ed8ff61","Type":"ContainerDied","Data":"2dde950ab7e904b314db9f7ca0bf17fd8d547f3d7bc69b805bec261f0aee8110"} Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.265202 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.284323 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" (UID: "f4f69ff2-19d7-4a2b-a34e-75062ed8ff61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.284677 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-config-data" (OuterVolumeSpecName: "config-data") pod "f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" (UID: "f4f69ff2-19d7-4a2b-a34e-75062ed8ff61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.285917 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-config-data" (OuterVolumeSpecName: "config-data") pod "04df1809-4543-4ca9-984f-d3547cbff9d4" (UID: "04df1809-4543-4ca9-984f-d3547cbff9d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.295377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04df1809-4543-4ca9-984f-d3547cbff9d4" (UID: "04df1809-4543-4ca9-984f-d3547cbff9d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.298248 4824 scope.go:117] "RemoveContainer" containerID="0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795" Nov 24 13:37:04 crc kubenswrapper[4824]: E1124 13:37:04.298853 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795\": container with ID starting with 0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795 not found: ID does not exist" containerID="0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.298884 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795"} err="failed to get container status \"0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795\": rpc error: code = NotFound desc = could not find container \"0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795\": container with ID starting with 0d422f56a8f05328442b9a84435847e990ce3c5d081c5f938bea04857c9f4795 not found: ID does not exist" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.298914 4824 scope.go:117] "RemoveContainer" containerID="543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.329017 4824 scope.go:117] "RemoveContainer" containerID="5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.348573 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c67t6\" (UniqueName: \"kubernetes.io/projected/04df1809-4543-4ca9-984f-d3547cbff9d4-kube-api-access-c67t6\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.348603 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.348614 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.348621 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.348630 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.348638 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04df1809-4543-4ca9-984f-d3547cbff9d4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.348645 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpsxw\" (UniqueName: \"kubernetes.io/projected/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61-kube-api-access-fpsxw\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.353757 4824 scope.go:117] "RemoveContainer" containerID="543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f" Nov 24 13:37:04 crc kubenswrapper[4824]: E1124 13:37:04.354229 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f\": container with ID starting with 543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f not found: ID does not exist" containerID="543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.354265 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f"} err="failed to get container status \"543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f\": rpc error: code = NotFound desc = could not find container \"543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f\": container with ID starting with 543b03e25b561b48cdae1cf9bcfe133b502c34e84f583ff8b424268b3fb86c5f not found: ID does not exist" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.354285 4824 scope.go:117] "RemoveContainer" containerID="5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa" Nov 24 13:37:04 crc kubenswrapper[4824]: E1124 13:37:04.354572 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa\": container with ID starting with 5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa not found: ID does not exist" containerID="5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.354601 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa"} err="failed to get container status \"5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa\": rpc error: code = NotFound desc = could not find container \"5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa\": container with ID starting with 5fa429b1fd739af8c53148e695668ae75c4a4428f995aa707a28ce5e6e1f39fa not found: ID does not exist" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.605634 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.613782 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.631542 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.645487 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.669889 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:37:04 crc kubenswrapper[4824]: E1124 13:37:04.670373 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerName="nova-metadata-log" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.670394 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerName="nova-metadata-log" Nov 24 13:37:04 crc kubenswrapper[4824]: E1124 13:37:04.670440 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerName="nova-metadata-metadata" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.670449 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerName="nova-metadata-metadata" Nov 24 13:37:04 crc kubenswrapper[4824]: E1124 13:37:04.670477 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04df1809-4543-4ca9-984f-d3547cbff9d4" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.670489 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="04df1809-4543-4ca9-984f-d3547cbff9d4" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.670704 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerName="nova-metadata-log" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.670728 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" containerName="nova-metadata-metadata" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.670739 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="04df1809-4543-4ca9-984f-d3547cbff9d4" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.671539 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.680392 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.680649 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.680818 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.682410 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.682852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.694429 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.707478 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.707666 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.720914 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.753930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.754000 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7pqf\" (UniqueName: \"kubernetes.io/projected/9be10433-b401-4853-ad5a-f32b7668fde2-kube-api-access-d7pqf\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.754049 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-logs\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.754130 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsfzr\" (UniqueName: \"kubernetes.io/projected/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-kube-api-access-wsfzr\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.754169 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.754265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.754307 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.754330 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.754365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-config-data\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.754467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.856844 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.856938 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-config-data\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.857076 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.857213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.857302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7pqf\" (UniqueName: \"kubernetes.io/projected/9be10433-b401-4853-ad5a-f32b7668fde2-kube-api-access-d7pqf\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.857397 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-logs\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.857645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsfzr\" (UniqueName: \"kubernetes.io/projected/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-kube-api-access-wsfzr\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.857746 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.857884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.857965 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.858434 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-logs\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.862366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-config-data\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.863084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.863194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.865630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.867911 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.868976 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.872284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9be10433-b401-4853-ad5a-f32b7668fde2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.883083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7pqf\" (UniqueName: \"kubernetes.io/projected/9be10433-b401-4853-ad5a-f32b7668fde2-kube-api-access-d7pqf\") pod \"nova-cell1-novncproxy-0\" (UID: \"9be10433-b401-4853-ad5a-f32b7668fde2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:04 crc kubenswrapper[4824]: I1124 13:37:04.894287 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsfzr\" (UniqueName: \"kubernetes.io/projected/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-kube-api-access-wsfzr\") pod \"nova-metadata-0\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " pod="openstack/nova-metadata-0" Nov 24 13:37:05 crc kubenswrapper[4824]: I1124 13:37:05.022586 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04df1809-4543-4ca9-984f-d3547cbff9d4" path="/var/lib/kubelet/pods/04df1809-4543-4ca9-984f-d3547cbff9d4/volumes" Nov 24 13:37:05 crc kubenswrapper[4824]: I1124 13:37:05.023330 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4f69ff2-19d7-4a2b-a34e-75062ed8ff61" path="/var/lib/kubelet/pods/f4f69ff2-19d7-4a2b-a34e-75062ed8ff61/volumes" Nov 24 13:37:05 crc kubenswrapper[4824]: I1124 13:37:05.042073 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:05 crc kubenswrapper[4824]: I1124 13:37:05.049687 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:05 crc kubenswrapper[4824]: I1124 13:37:05.512045 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:05 crc kubenswrapper[4824]: W1124 13:37:05.513457 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc6a6b8f_d90e_4546_be92_3bd14fb9bc4f.slice/crio-6451790550c7ecd6971d2fc31ccd7fdde3a67a2fa94f1e4482149e222559eec1 WatchSource:0}: Error finding container 6451790550c7ecd6971d2fc31ccd7fdde3a67a2fa94f1e4482149e222559eec1: Status 404 returned error can't find the container with id 6451790550c7ecd6971d2fc31ccd7fdde3a67a2fa94f1e4482149e222559eec1 Nov 24 13:37:05 crc kubenswrapper[4824]: I1124 13:37:05.574702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 13:37:05 crc kubenswrapper[4824]: W1124 13:37:05.583867 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9be10433_b401_4853_ad5a_f32b7668fde2.slice/crio-deef27c46e43ea117b0d81ca9efc896995cf06c121eec4f46dea436fffafcc1e WatchSource:0}: Error finding container deef27c46e43ea117b0d81ca9efc896995cf06c121eec4f46dea436fffafcc1e: Status 404 returned error can't find the container with id deef27c46e43ea117b0d81ca9efc896995cf06c121eec4f46dea436fffafcc1e Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.228018 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.228675 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.230091 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.234305 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.286746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f","Type":"ContainerStarted","Data":"52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75"} Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.286789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f","Type":"ContainerStarted","Data":"e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6"} Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.286799 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f","Type":"ContainerStarted","Data":"6451790550c7ecd6971d2fc31ccd7fdde3a67a2fa94f1e4482149e222559eec1"} Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.288973 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9be10433-b401-4853-ad5a-f32b7668fde2","Type":"ContainerStarted","Data":"aebf8436202f710b2d9fe7b0d556bb46dee80efd59905a7fdd0bbda3b889fc55"} Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.289007 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9be10433-b401-4853-ad5a-f32b7668fde2","Type":"ContainerStarted","Data":"deef27c46e43ea117b0d81ca9efc896995cf06c121eec4f46dea436fffafcc1e"} Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.289155 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.292616 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.309125 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.309101896 podStartE2EDuration="2.309101896s" podCreationTimestamp="2025-11-24 13:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:37:06.304404253 +0000 UTC m=+1367.943943563" watchObservedRunningTime="2025-11-24 13:37:06.309101896 +0000 UTC m=+1367.948641216" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.326946 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.326927074 podStartE2EDuration="2.326927074s" podCreationTimestamp="2025-11-24 13:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:37:06.324044111 +0000 UTC m=+1367.963583421" watchObservedRunningTime="2025-11-24 13:37:06.326927074 +0000 UTC m=+1367.966466384" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.529079 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6g4tc"] Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.531057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.546958 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6g4tc"] Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.592027 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.592178 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.592239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.592258 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-config\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.592286 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf5qn\" (UniqueName: \"kubernetes.io/projected/a2efd9de-42a7-49c7-a528-54919f22fbea-kube-api-access-tf5qn\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.592361 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.693507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.693678 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.693796 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.693842 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.693857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-config\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.693882 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf5qn\" (UniqueName: \"kubernetes.io/projected/a2efd9de-42a7-49c7-a528-54919f22fbea-kube-api-access-tf5qn\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.695477 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.695501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.696098 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-config\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.697365 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.698298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.717418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf5qn\" (UniqueName: \"kubernetes.io/projected/a2efd9de-42a7-49c7-a528-54919f22fbea-kube-api-access-tf5qn\") pod \"dnsmasq-dns-89c5cd4d5-6g4tc\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:06 crc kubenswrapper[4824]: I1124 13:37:06.861090 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:07 crc kubenswrapper[4824]: I1124 13:37:07.348696 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6g4tc"] Nov 24 13:37:08 crc kubenswrapper[4824]: I1124 13:37:08.309919 4824 generic.go:334] "Generic (PLEG): container finished" podID="a2efd9de-42a7-49c7-a528-54919f22fbea" containerID="cb621c6b650ce5f8840fec7b5846413f285f346f675c4126d916f68833f321e6" exitCode=0 Nov 24 13:37:08 crc kubenswrapper[4824]: I1124 13:37:08.311000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" event={"ID":"a2efd9de-42a7-49c7-a528-54919f22fbea","Type":"ContainerDied","Data":"cb621c6b650ce5f8840fec7b5846413f285f346f675c4126d916f68833f321e6"} Nov 24 13:37:08 crc kubenswrapper[4824]: I1124 13:37:08.311138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" event={"ID":"a2efd9de-42a7-49c7-a528-54919f22fbea","Type":"ContainerStarted","Data":"0659e61b7b260b13491ff10a57334371321c8b9f4636b720eff8591b6b4c1f56"} Nov 24 13:37:08 crc kubenswrapper[4824]: I1124 13:37:08.940135 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:37:08 crc kubenswrapper[4824]: I1124 13:37:08.941129 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="ceilometer-central-agent" containerID="cri-o://906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049" gracePeriod=30 Nov 24 13:37:08 crc kubenswrapper[4824]: I1124 13:37:08.941242 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="sg-core" containerID="cri-o://d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f" gracePeriod=30 Nov 24 13:37:08 crc kubenswrapper[4824]: I1124 13:37:08.941435 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="proxy-httpd" containerID="cri-o://7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea" gracePeriod=30 Nov 24 13:37:08 crc kubenswrapper[4824]: I1124 13:37:08.941516 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="ceilometer-notification-agent" containerID="cri-o://969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073" gracePeriod=30 Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.041680 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.321241 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" event={"ID":"a2efd9de-42a7-49c7-a528-54919f22fbea","Type":"ContainerStarted","Data":"25712dc18d9ce8eec007b3fbc507442babc42d3505dbc0dd5f8b31192ea2a0e0"} Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.321365 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.326514 4824 generic.go:334] "Generic (PLEG): container finished" podID="d83e37fd-645e-4258-a441-a9a578844485" containerID="7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea" exitCode=0 Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.326541 4824 generic.go:334] "Generic (PLEG): container finished" podID="d83e37fd-645e-4258-a441-a9a578844485" containerID="d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f" exitCode=2 Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.326561 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerDied","Data":"7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea"} Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.326625 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerDied","Data":"d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f"} Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.326722 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-log" containerID="cri-o://2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac" gracePeriod=30 Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.326838 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-api" containerID="cri-o://edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd" gracePeriod=30 Nov 24 13:37:09 crc kubenswrapper[4824]: I1124 13:37:09.359491 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" podStartSLOduration=3.359472291 podStartE2EDuration="3.359472291s" podCreationTimestamp="2025-11-24 13:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:37:09.355214818 +0000 UTC m=+1370.994754128" watchObservedRunningTime="2025-11-24 13:37:09.359472291 +0000 UTC m=+1370.999011601" Nov 24 13:37:10 crc kubenswrapper[4824]: I1124 13:37:10.042266 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:10 crc kubenswrapper[4824]: I1124 13:37:10.050961 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:37:10 crc kubenswrapper[4824]: I1124 13:37:10.051007 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:37:10 crc kubenswrapper[4824]: I1124 13:37:10.369330 4824 generic.go:334] "Generic (PLEG): container finished" podID="d83e37fd-645e-4258-a441-a9a578844485" containerID="906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049" exitCode=0 Nov 24 13:37:10 crc kubenswrapper[4824]: I1124 13:37:10.371905 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerDied","Data":"906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049"} Nov 24 13:37:10 crc kubenswrapper[4824]: I1124 13:37:10.393309 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerID="2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac" exitCode=143 Nov 24 13:37:10 crc kubenswrapper[4824]: I1124 13:37:10.394066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f2c2432-5906-40b9-b6ff-795783620bc2","Type":"ContainerDied","Data":"2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac"} Nov 24 13:37:12 crc kubenswrapper[4824]: I1124 13:37:12.915211 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.045891 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrbvk\" (UniqueName: \"kubernetes.io/projected/9f2c2432-5906-40b9-b6ff-795783620bc2-kube-api-access-xrbvk\") pod \"9f2c2432-5906-40b9-b6ff-795783620bc2\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.046512 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f2c2432-5906-40b9-b6ff-795783620bc2-logs\") pod \"9f2c2432-5906-40b9-b6ff-795783620bc2\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.046616 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-combined-ca-bundle\") pod \"9f2c2432-5906-40b9-b6ff-795783620bc2\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.046758 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-config-data\") pod \"9f2c2432-5906-40b9-b6ff-795783620bc2\" (UID: \"9f2c2432-5906-40b9-b6ff-795783620bc2\") " Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.048039 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f2c2432-5906-40b9-b6ff-795783620bc2-logs" (OuterVolumeSpecName: "logs") pod "9f2c2432-5906-40b9-b6ff-795783620bc2" (UID: "9f2c2432-5906-40b9-b6ff-795783620bc2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.065071 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f2c2432-5906-40b9-b6ff-795783620bc2-kube-api-access-xrbvk" (OuterVolumeSpecName: "kube-api-access-xrbvk") pod "9f2c2432-5906-40b9-b6ff-795783620bc2" (UID: "9f2c2432-5906-40b9-b6ff-795783620bc2"). InnerVolumeSpecName "kube-api-access-xrbvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.090302 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f2c2432-5906-40b9-b6ff-795783620bc2" (UID: "9f2c2432-5906-40b9-b6ff-795783620bc2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.111998 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-config-data" (OuterVolumeSpecName: "config-data") pod "9f2c2432-5906-40b9-b6ff-795783620bc2" (UID: "9f2c2432-5906-40b9-b6ff-795783620bc2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.149576 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrbvk\" (UniqueName: \"kubernetes.io/projected/9f2c2432-5906-40b9-b6ff-795783620bc2-kube-api-access-xrbvk\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.149613 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f2c2432-5906-40b9-b6ff-795783620bc2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.149622 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.149631 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f2c2432-5906-40b9-b6ff-795783620bc2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.425442 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerID="edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd" exitCode=0 Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.425488 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f2c2432-5906-40b9-b6ff-795783620bc2","Type":"ContainerDied","Data":"edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd"} Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.425524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f2c2432-5906-40b9-b6ff-795783620bc2","Type":"ContainerDied","Data":"5273f19b62974ab23ac94ddd12bb7cebea98e515cec50a990f5641d8656acbd0"} Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.425545 4824 scope.go:117] "RemoveContainer" containerID="edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.425599 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.466324 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.471609 4824 scope.go:117] "RemoveContainer" containerID="2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.480750 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.500127 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:13 crc kubenswrapper[4824]: E1124 13:37:13.500604 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-log" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.500627 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-log" Nov 24 13:37:13 crc kubenswrapper[4824]: E1124 13:37:13.500639 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-api" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.500645 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-api" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.500858 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-log" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.500869 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" containerName="nova-api-api" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.501823 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.507753 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.508070 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.508163 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.531547 4824 scope.go:117] "RemoveContainer" containerID="edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.535795 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:13 crc kubenswrapper[4824]: E1124 13:37:13.544102 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd\": container with ID starting with edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd not found: ID does not exist" containerID="edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.544148 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd"} err="failed to get container status \"edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd\": rpc error: code = NotFound desc = could not find container \"edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd\": container with ID starting with edddc0198cf96ffdb55dbcac14c5e0f84daa1f48cb24fd1532ec6d5d2551c9fd not found: ID does not exist" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.544177 4824 scope.go:117] "RemoveContainer" containerID="2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac" Nov 24 13:37:13 crc kubenswrapper[4824]: E1124 13:37:13.544860 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac\": container with ID starting with 2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac not found: ID does not exist" containerID="2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.545022 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac"} err="failed to get container status \"2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac\": rpc error: code = NotFound desc = could not find container \"2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac\": container with ID starting with 2b7f39b336f90bbea5de56d38b6ac227eaf3272b5a5323e5a9c2d1e75b68dcac not found: ID does not exist" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.658925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.659204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-public-tls-certs\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.659318 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-config-data\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.659392 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7972c42a-2af1-4d7c-81a6-cdc342978f8b-logs\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.659483 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rllzq\" (UniqueName: \"kubernetes.io/projected/7972c42a-2af1-4d7c-81a6-cdc342978f8b-kube-api-access-rllzq\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.659591 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.761629 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.762094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.762165 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-public-tls-certs\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.762246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-config-data\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.762269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7972c42a-2af1-4d7c-81a6-cdc342978f8b-logs\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.762328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rllzq\" (UniqueName: \"kubernetes.io/projected/7972c42a-2af1-4d7c-81a6-cdc342978f8b-kube-api-access-rllzq\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.762953 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7972c42a-2af1-4d7c-81a6-cdc342978f8b-logs\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.768247 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.770072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-public-tls-certs\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.781596 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-config-data\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.783462 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.786919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rllzq\" (UniqueName: \"kubernetes.io/projected/7972c42a-2af1-4d7c-81a6-cdc342978f8b-kube-api-access-rllzq\") pod \"nova-api-0\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " pod="openstack/nova-api-0" Nov 24 13:37:13 crc kubenswrapper[4824]: I1124 13:37:13.828402 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:14 crc kubenswrapper[4824]: I1124 13:37:14.329102 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:14 crc kubenswrapper[4824]: I1124 13:37:14.434956 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7972c42a-2af1-4d7c-81a6-cdc342978f8b","Type":"ContainerStarted","Data":"1be3a431d1a61341162d7332e376e0c853cd9d4bbc0a2f00fce98e5ba2edc9d5"} Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.020237 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f2c2432-5906-40b9-b6ff-795783620bc2" path="/var/lib/kubelet/pods/9f2c2432-5906-40b9-b6ff-795783620bc2/volumes" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.042258 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.050559 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.050600 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.062329 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.114247 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.290783 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-ceilometer-tls-certs\") pod \"d83e37fd-645e-4258-a441-a9a578844485\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.290946 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-config-data\") pod \"d83e37fd-645e-4258-a441-a9a578844485\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.290968 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-log-httpd\") pod \"d83e37fd-645e-4258-a441-a9a578844485\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.290995 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-combined-ca-bundle\") pod \"d83e37fd-645e-4258-a441-a9a578844485\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.291032 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbxsq\" (UniqueName: \"kubernetes.io/projected/d83e37fd-645e-4258-a441-a9a578844485-kube-api-access-rbxsq\") pod \"d83e37fd-645e-4258-a441-a9a578844485\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.291103 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-run-httpd\") pod \"d83e37fd-645e-4258-a441-a9a578844485\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.291126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-scripts\") pod \"d83e37fd-645e-4258-a441-a9a578844485\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.291163 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-sg-core-conf-yaml\") pod \"d83e37fd-645e-4258-a441-a9a578844485\" (UID: \"d83e37fd-645e-4258-a441-a9a578844485\") " Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.291365 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d83e37fd-645e-4258-a441-a9a578844485" (UID: "d83e37fd-645e-4258-a441-a9a578844485"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.291862 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.292061 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d83e37fd-645e-4258-a441-a9a578844485" (UID: "d83e37fd-645e-4258-a441-a9a578844485"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.300246 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-scripts" (OuterVolumeSpecName: "scripts") pod "d83e37fd-645e-4258-a441-a9a578844485" (UID: "d83e37fd-645e-4258-a441-a9a578844485"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.303014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d83e37fd-645e-4258-a441-a9a578844485-kube-api-access-rbxsq" (OuterVolumeSpecName: "kube-api-access-rbxsq") pod "d83e37fd-645e-4258-a441-a9a578844485" (UID: "d83e37fd-645e-4258-a441-a9a578844485"). InnerVolumeSpecName "kube-api-access-rbxsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.330096 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d83e37fd-645e-4258-a441-a9a578844485" (UID: "d83e37fd-645e-4258-a441-a9a578844485"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.358599 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d83e37fd-645e-4258-a441-a9a578844485" (UID: "d83e37fd-645e-4258-a441-a9a578844485"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.382616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d83e37fd-645e-4258-a441-a9a578844485" (UID: "d83e37fd-645e-4258-a441-a9a578844485"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.393844 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.393875 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbxsq\" (UniqueName: \"kubernetes.io/projected/d83e37fd-645e-4258-a441-a9a578844485-kube-api-access-rbxsq\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.393887 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d83e37fd-645e-4258-a441-a9a578844485-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.393894 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.393903 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.393911 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.404057 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-config-data" (OuterVolumeSpecName: "config-data") pod "d83e37fd-645e-4258-a441-a9a578844485" (UID: "d83e37fd-645e-4258-a441-a9a578844485"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.451822 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7972c42a-2af1-4d7c-81a6-cdc342978f8b","Type":"ContainerStarted","Data":"b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e"} Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.451882 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7972c42a-2af1-4d7c-81a6-cdc342978f8b","Type":"ContainerStarted","Data":"ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7"} Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.454771 4824 generic.go:334] "Generic (PLEG): container finished" podID="d83e37fd-645e-4258-a441-a9a578844485" containerID="969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073" exitCode=0 Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.455741 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.455752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerDied","Data":"969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073"} Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.455971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d83e37fd-645e-4258-a441-a9a578844485","Type":"ContainerDied","Data":"d15f2f855052d5a210caca2e9c6174bcc57899c2489235623a54239e235c008f"} Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.455992 4824 scope.go:117] "RemoveContainer" containerID="7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.478891 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.47886944 podStartE2EDuration="2.47886944s" podCreationTimestamp="2025-11-24 13:37:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:37:15.473407821 +0000 UTC m=+1377.112947131" watchObservedRunningTime="2025-11-24 13:37:15.47886944 +0000 UTC m=+1377.118408750" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.489871 4824 scope.go:117] "RemoveContainer" containerID="d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.495464 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83e37fd-645e-4258-a441-a9a578844485-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.505218 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.516386 4824 scope.go:117] "RemoveContainer" containerID="969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.524917 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.534910 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.557111 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:37:15 crc kubenswrapper[4824]: E1124 13:37:15.557584 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="ceilometer-notification-agent" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.557601 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="ceilometer-notification-agent" Nov 24 13:37:15 crc kubenswrapper[4824]: E1124 13:37:15.557615 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="sg-core" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.557622 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="sg-core" Nov 24 13:37:15 crc kubenswrapper[4824]: E1124 13:37:15.557633 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="ceilometer-central-agent" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.557641 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="ceilometer-central-agent" Nov 24 13:37:15 crc kubenswrapper[4824]: E1124 13:37:15.557660 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="proxy-httpd" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.557666 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="proxy-httpd" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.557857 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="sg-core" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.557872 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="ceilometer-notification-agent" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.557891 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="proxy-httpd" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.557902 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d83e37fd-645e-4258-a441-a9a578844485" containerName="ceilometer-central-agent" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.559579 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.565408 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.565754 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.568451 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.588299 4824 scope.go:117] "RemoveContainer" containerID="906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.619103 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.658958 4824 scope.go:117] "RemoveContainer" containerID="7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea" Nov 24 13:37:15 crc kubenswrapper[4824]: E1124 13:37:15.662915 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea\": container with ID starting with 7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea not found: ID does not exist" containerID="7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.662954 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea"} err="failed to get container status \"7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea\": rpc error: code = NotFound desc = could not find container \"7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea\": container with ID starting with 7f6a3c81d1ad41ef21989199ae82e6c89d2747d363a627365a75921822637aea not found: ID does not exist" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.662976 4824 scope.go:117] "RemoveContainer" containerID="d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f" Nov 24 13:37:15 crc kubenswrapper[4824]: E1124 13:37:15.663278 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f\": container with ID starting with d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f not found: ID does not exist" containerID="d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.663299 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f"} err="failed to get container status \"d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f\": rpc error: code = NotFound desc = could not find container \"d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f\": container with ID starting with d1b43fa5a9b148cfddb228961ba2d734dc2f96b292b37bfe6383243ba0e7cf5f not found: ID does not exist" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.663312 4824 scope.go:117] "RemoveContainer" containerID="969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073" Nov 24 13:37:15 crc kubenswrapper[4824]: E1124 13:37:15.663606 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073\": container with ID starting with 969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073 not found: ID does not exist" containerID="969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.663644 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073"} err="failed to get container status \"969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073\": rpc error: code = NotFound desc = could not find container \"969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073\": container with ID starting with 969bfa1f07385d05d32cc095b9c3041ae33d94244c1d2d5c1ebe129ea7300073 not found: ID does not exist" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.663672 4824 scope.go:117] "RemoveContainer" containerID="906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049" Nov 24 13:37:15 crc kubenswrapper[4824]: E1124 13:37:15.664728 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049\": container with ID starting with 906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049 not found: ID does not exist" containerID="906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.664748 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049"} err="failed to get container status \"906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049\": rpc error: code = NotFound desc = could not find container \"906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049\": container with ID starting with 906add1d83f3bf0be8acfb8c4491bf0576674c5301b492b414d21cc7b8cb3049 not found: ID does not exist" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.704870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f521f794-dddf-4c12-9b64-665b541d43ec-log-httpd\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.704947 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwcqp\" (UniqueName: \"kubernetes.io/projected/f521f794-dddf-4c12-9b64-665b541d43ec-kube-api-access-lwcqp\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.704975 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.705006 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f521f794-dddf-4c12-9b64-665b541d43ec-run-httpd\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.705024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-scripts\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.705037 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.705085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.705126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-config-data\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.728876 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-q9524"] Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.730386 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.736953 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.737360 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.747117 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-q9524"] Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.807034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-config-data\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.807135 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f521f794-dddf-4c12-9b64-665b541d43ec-log-httpd\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.807239 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwcqp\" (UniqueName: \"kubernetes.io/projected/f521f794-dddf-4c12-9b64-665b541d43ec-kube-api-access-lwcqp\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.808007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f521f794-dddf-4c12-9b64-665b541d43ec-log-httpd\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.808085 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.808146 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f521f794-dddf-4c12-9b64-665b541d43ec-run-httpd\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.808172 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-scripts\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.808192 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.808292 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.808648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f521f794-dddf-4c12-9b64-665b541d43ec-run-httpd\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.811637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.812538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-config-data\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.820761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-scripts\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.821474 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.826640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f521f794-dddf-4c12-9b64-665b541d43ec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.827002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwcqp\" (UniqueName: \"kubernetes.io/projected/f521f794-dddf-4c12-9b64-665b541d43ec-kube-api-access-lwcqp\") pod \"ceilometer-0\" (UID: \"f521f794-dddf-4c12-9b64-665b541d43ec\") " pod="openstack/ceilometer-0" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.909902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.909983 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-scripts\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.910072 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-config-data\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.910111 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc8g6\" (UniqueName: \"kubernetes.io/projected/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-kube-api-access-rc8g6\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:15 crc kubenswrapper[4824]: I1124 13:37:15.953336 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.011329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-config-data\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.011390 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc8g6\" (UniqueName: \"kubernetes.io/projected/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-kube-api-access-rc8g6\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.011417 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.011467 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-scripts\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.015489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.016146 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-scripts\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.016345 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-config-data\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.030854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc8g6\" (UniqueName: \"kubernetes.io/projected/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-kube-api-access-rc8g6\") pod \"nova-cell1-cell-mapping-q9524\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.054156 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.063018 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.063061 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.443993 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 13:37:16 crc kubenswrapper[4824]: W1124 13:37:16.444165 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf521f794_dddf_4c12_9b64_665b541d43ec.slice/crio-180c2a07a4b5e542ceaac67a305b04af2a838c11999604805394afdb31e232b4 WatchSource:0}: Error finding container 180c2a07a4b5e542ceaac67a305b04af2a838c11999604805394afdb31e232b4: Status 404 returned error can't find the container with id 180c2a07a4b5e542ceaac67a305b04af2a838c11999604805394afdb31e232b4 Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.463693 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f521f794-dddf-4c12-9b64-665b541d43ec","Type":"ContainerStarted","Data":"180c2a07a4b5e542ceaac67a305b04af2a838c11999604805394afdb31e232b4"} Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.589705 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-q9524"] Nov 24 13:37:16 crc kubenswrapper[4824]: W1124 13:37:16.598238 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c284e54_883c_4ea7_9ce2_bc6bc4a9c8cf.slice/crio-f2826566ba65f95a2325a04baa6a8961f6af049ba1d3180b88a7377c321d25f7 WatchSource:0}: Error finding container f2826566ba65f95a2325a04baa6a8961f6af049ba1d3180b88a7377c321d25f7: Status 404 returned error can't find the container with id f2826566ba65f95a2325a04baa6a8961f6af049ba1d3180b88a7377c321d25f7 Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.864475 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.951323 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kgmm4"] Nov 24 13:37:16 crc kubenswrapper[4824]: I1124 13:37:16.952058 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" podUID="1a630e6e-19e4-424c-a393-2f08c2c5a154" containerName="dnsmasq-dns" containerID="cri-o://8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447" gracePeriod=10 Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.026862 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d83e37fd-645e-4258-a441-a9a578844485" path="/var/lib/kubelet/pods/d83e37fd-645e-4258-a441-a9a578844485/volumes" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.493476 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.494981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f521f794-dddf-4c12-9b64-665b541d43ec","Type":"ContainerStarted","Data":"6dae4d2ee121fcc39c7ed6e470bd32a3afce37db51d6a340956c8c90d315b2ae"} Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.500300 4824 generic.go:334] "Generic (PLEG): container finished" podID="1a630e6e-19e4-424c-a393-2f08c2c5a154" containerID="8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447" exitCode=0 Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.500356 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" event={"ID":"1a630e6e-19e4-424c-a393-2f08c2c5a154","Type":"ContainerDied","Data":"8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447"} Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.500380 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" event={"ID":"1a630e6e-19e4-424c-a393-2f08c2c5a154","Type":"ContainerDied","Data":"495861b584582646c936fa48b1def5a265eb9e674ba0f9642c4976e713caa9f3"} Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.500395 4824 scope.go:117] "RemoveContainer" containerID="8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.500505 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-kgmm4" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.512001 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q9524" event={"ID":"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf","Type":"ContainerStarted","Data":"b6ba79778b2ba939a1dca97c40ae09b5ac5f5f75367717e7a9caa7280cca9c94"} Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.512034 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q9524" event={"ID":"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf","Type":"ContainerStarted","Data":"f2826566ba65f95a2325a04baa6a8961f6af049ba1d3180b88a7377c321d25f7"} Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.564311 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-q9524" podStartSLOduration=2.564291985 podStartE2EDuration="2.564291985s" podCreationTimestamp="2025-11-24 13:37:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:37:17.557691012 +0000 UTC m=+1379.197230322" watchObservedRunningTime="2025-11-24 13:37:17.564291985 +0000 UTC m=+1379.203831295" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.596639 4824 scope.go:117] "RemoveContainer" containerID="545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.631989 4824 scope.go:117] "RemoveContainer" containerID="8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447" Nov 24 13:37:17 crc kubenswrapper[4824]: E1124 13:37:17.634473 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447\": container with ID starting with 8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447 not found: ID does not exist" containerID="8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.634519 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447"} err="failed to get container status \"8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447\": rpc error: code = NotFound desc = could not find container \"8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447\": container with ID starting with 8f8ab7fbd91db1c27cf3dcff81507bbb218f9bb8119a9cb9a815ef8f578bb447 not found: ID does not exist" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.634548 4824 scope.go:117] "RemoveContainer" containerID="545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5" Nov 24 13:37:17 crc kubenswrapper[4824]: E1124 13:37:17.635130 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5\": container with ID starting with 545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5 not found: ID does not exist" containerID="545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.635171 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5"} err="failed to get container status \"545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5\": rpc error: code = NotFound desc = could not find container \"545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5\": container with ID starting with 545853061b079f05202b2dd7662813a4b3bef8c9c4cc01d3b365f2880266d8a5 not found: ID does not exist" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.658424 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-sb\") pod \"1a630e6e-19e4-424c-a393-2f08c2c5a154\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.658566 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nv78\" (UniqueName: \"kubernetes.io/projected/1a630e6e-19e4-424c-a393-2f08c2c5a154-kube-api-access-8nv78\") pod \"1a630e6e-19e4-424c-a393-2f08c2c5a154\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.658618 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-nb\") pod \"1a630e6e-19e4-424c-a393-2f08c2c5a154\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.658660 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-config\") pod \"1a630e6e-19e4-424c-a393-2f08c2c5a154\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.660057 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-swift-storage-0\") pod \"1a630e6e-19e4-424c-a393-2f08c2c5a154\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.660096 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-svc\") pod \"1a630e6e-19e4-424c-a393-2f08c2c5a154\" (UID: \"1a630e6e-19e4-424c-a393-2f08c2c5a154\") " Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.679154 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a630e6e-19e4-424c-a393-2f08c2c5a154-kube-api-access-8nv78" (OuterVolumeSpecName: "kube-api-access-8nv78") pod "1a630e6e-19e4-424c-a393-2f08c2c5a154" (UID: "1a630e6e-19e4-424c-a393-2f08c2c5a154"). InnerVolumeSpecName "kube-api-access-8nv78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.731545 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1a630e6e-19e4-424c-a393-2f08c2c5a154" (UID: "1a630e6e-19e4-424c-a393-2f08c2c5a154"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.762543 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nv78\" (UniqueName: \"kubernetes.io/projected/1a630e6e-19e4-424c-a393-2f08c2c5a154-kube-api-access-8nv78\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.762587 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.770348 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1a630e6e-19e4-424c-a393-2f08c2c5a154" (UID: "1a630e6e-19e4-424c-a393-2f08c2c5a154"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.772049 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-config" (OuterVolumeSpecName: "config") pod "1a630e6e-19e4-424c-a393-2f08c2c5a154" (UID: "1a630e6e-19e4-424c-a393-2f08c2c5a154"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.776525 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1a630e6e-19e4-424c-a393-2f08c2c5a154" (UID: "1a630e6e-19e4-424c-a393-2f08c2c5a154"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.818249 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1a630e6e-19e4-424c-a393-2f08c2c5a154" (UID: "1a630e6e-19e4-424c-a393-2f08c2c5a154"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.864782 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.864844 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.864861 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:17 crc kubenswrapper[4824]: I1124 13:37:17.864873 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a630e6e-19e4-424c-a393-2f08c2c5a154-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:18 crc kubenswrapper[4824]: I1124 13:37:18.139454 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kgmm4"] Nov 24 13:37:18 crc kubenswrapper[4824]: I1124 13:37:18.147079 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kgmm4"] Nov 24 13:37:18 crc kubenswrapper[4824]: I1124 13:37:18.521639 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f521f794-dddf-4c12-9b64-665b541d43ec","Type":"ContainerStarted","Data":"1e95ab8909f7a8300f1b0740891bf130de376cb914f3586ffc02f13f6d0fc51a"} Nov 24 13:37:19 crc kubenswrapper[4824]: I1124 13:37:19.025270 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a630e6e-19e4-424c-a393-2f08c2c5a154" path="/var/lib/kubelet/pods/1a630e6e-19e4-424c-a393-2f08c2c5a154/volumes" Nov 24 13:37:19 crc kubenswrapper[4824]: I1124 13:37:19.538921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f521f794-dddf-4c12-9b64-665b541d43ec","Type":"ContainerStarted","Data":"773a1826fb8f9c0dcad2ddc88072e2470b239103227c59c2cb7870f5712cbcd8"} Nov 24 13:37:20 crc kubenswrapper[4824]: I1124 13:37:20.558794 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f521f794-dddf-4c12-9b64-665b541d43ec","Type":"ContainerStarted","Data":"f74a66c20a31c539dbe8ff6cee9e55ce60fed92547649033a1c15e017dbc6e72"} Nov 24 13:37:20 crc kubenswrapper[4824]: I1124 13:37:20.559220 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 13:37:20 crc kubenswrapper[4824]: I1124 13:37:20.583036 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9728190209999998 podStartE2EDuration="5.583017682s" podCreationTimestamp="2025-11-24 13:37:15 +0000 UTC" firstStartedPulling="2025-11-24 13:37:16.4489731 +0000 UTC m=+1378.088512410" lastFinishedPulling="2025-11-24 13:37:20.059171761 +0000 UTC m=+1381.698711071" observedRunningTime="2025-11-24 13:37:20.581938308 +0000 UTC m=+1382.221477618" watchObservedRunningTime="2025-11-24 13:37:20.583017682 +0000 UTC m=+1382.222556992" Nov 24 13:37:22 crc kubenswrapper[4824]: E1124 13:37:22.363977 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c284e54_883c_4ea7_9ce2_bc6bc4a9c8cf.slice/crio-conmon-b6ba79778b2ba939a1dca97c40ae09b5ac5f5f75367717e7a9caa7280cca9c94.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c284e54_883c_4ea7_9ce2_bc6bc4a9c8cf.slice/crio-b6ba79778b2ba939a1dca97c40ae09b5ac5f5f75367717e7a9caa7280cca9c94.scope\": RecentStats: unable to find data in memory cache]" Nov 24 13:37:22 crc kubenswrapper[4824]: I1124 13:37:22.577511 4824 generic.go:334] "Generic (PLEG): container finished" podID="0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf" containerID="b6ba79778b2ba939a1dca97c40ae09b5ac5f5f75367717e7a9caa7280cca9c94" exitCode=0 Nov 24 13:37:22 crc kubenswrapper[4824]: I1124 13:37:22.577554 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q9524" event={"ID":"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf","Type":"ContainerDied","Data":"b6ba79778b2ba939a1dca97c40ae09b5ac5f5f75367717e7a9caa7280cca9c94"} Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.381829 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hgdqw"] Nov 24 13:37:23 crc kubenswrapper[4824]: E1124 13:37:23.392937 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a630e6e-19e4-424c-a393-2f08c2c5a154" containerName="init" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.393194 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a630e6e-19e4-424c-a393-2f08c2c5a154" containerName="init" Nov 24 13:37:23 crc kubenswrapper[4824]: E1124 13:37:23.393329 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a630e6e-19e4-424c-a393-2f08c2c5a154" containerName="dnsmasq-dns" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.395209 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a630e6e-19e4-424c-a393-2f08c2c5a154" containerName="dnsmasq-dns" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.395510 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a630e6e-19e4-424c-a393-2f08c2c5a154" containerName="dnsmasq-dns" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.396992 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hgdqw"] Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.397183 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.466190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-catalog-content\") pod \"redhat-marketplace-hgdqw\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.466332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k47hd\" (UniqueName: \"kubernetes.io/projected/a2a19c34-3ca4-478e-a544-d9a5e48abb42-kube-api-access-k47hd\") pod \"redhat-marketplace-hgdqw\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.466354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-utilities\") pod \"redhat-marketplace-hgdqw\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.568840 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-catalog-content\") pod \"redhat-marketplace-hgdqw\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.569306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k47hd\" (UniqueName: \"kubernetes.io/projected/a2a19c34-3ca4-478e-a544-d9a5e48abb42-kube-api-access-k47hd\") pod \"redhat-marketplace-hgdqw\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.569336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-utilities\") pod \"redhat-marketplace-hgdqw\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.569480 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-catalog-content\") pod \"redhat-marketplace-hgdqw\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.569781 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-utilities\") pod \"redhat-marketplace-hgdqw\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.595666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k47hd\" (UniqueName: \"kubernetes.io/projected/a2a19c34-3ca4-478e-a544-d9a5e48abb42-kube-api-access-k47hd\") pod \"redhat-marketplace-hgdqw\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.729747 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.830058 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:37:23 crc kubenswrapper[4824]: I1124 13:37:23.830306 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.067686 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.183608 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc8g6\" (UniqueName: \"kubernetes.io/projected/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-kube-api-access-rc8g6\") pod \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.183691 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-config-data\") pod \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.183716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-combined-ca-bundle\") pod \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.183795 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-scripts\") pod \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\" (UID: \"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf\") " Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.189209 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-scripts" (OuterVolumeSpecName: "scripts") pod "0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf" (UID: "0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.189329 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-kube-api-access-rc8g6" (OuterVolumeSpecName: "kube-api-access-rc8g6") pod "0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf" (UID: "0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf"). InnerVolumeSpecName "kube-api-access-rc8g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.211670 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-config-data" (OuterVolumeSpecName: "config-data") pod "0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf" (UID: "0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.234967 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hgdqw"] Nov 24 13:37:24 crc kubenswrapper[4824]: W1124 13:37:24.237737 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2a19c34_3ca4_478e_a544_d9a5e48abb42.slice/crio-ffc77ea27c83ed44df8b54ca8f45a072d415fa85f48295f8eb234dbded782efc WatchSource:0}: Error finding container ffc77ea27c83ed44df8b54ca8f45a072d415fa85f48295f8eb234dbded782efc: Status 404 returned error can't find the container with id ffc77ea27c83ed44df8b54ca8f45a072d415fa85f48295f8eb234dbded782efc Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.238034 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf" (UID: "0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.286303 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc8g6\" (UniqueName: \"kubernetes.io/projected/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-kube-api-access-rc8g6\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.286559 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.286648 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.286722 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.605606 4824 generic.go:334] "Generic (PLEG): container finished" podID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerID="71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232" exitCode=0 Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.605966 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hgdqw" event={"ID":"a2a19c34-3ca4-478e-a544-d9a5e48abb42","Type":"ContainerDied","Data":"71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232"} Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.605997 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hgdqw" event={"ID":"a2a19c34-3ca4-478e-a544-d9a5e48abb42","Type":"ContainerStarted","Data":"ffc77ea27c83ed44df8b54ca8f45a072d415fa85f48295f8eb234dbded782efc"} Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.644738 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q9524" event={"ID":"0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf","Type":"ContainerDied","Data":"f2826566ba65f95a2325a04baa6a8961f6af049ba1d3180b88a7377c321d25f7"} Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.644792 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2826566ba65f95a2325a04baa6a8961f6af049ba1d3180b88a7377c321d25f7" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.644873 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q9524" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.783220 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.783460 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9661f7ee-75ed-4c09-be2d-d08d2c8eda1b" containerName="nova-scheduler-scheduler" containerID="cri-o://fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39" gracePeriod=30 Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.802082 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.802371 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-log" containerID="cri-o://ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7" gracePeriod=30 Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.802833 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-api" containerID="cri-o://b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e" gracePeriod=30 Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.815615 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.815896 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-log" containerID="cri-o://e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6" gracePeriod=30 Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.816045 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-metadata" containerID="cri-o://52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75" gracePeriod=30 Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.818554 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": EOF" Nov 24 13:37:24 crc kubenswrapper[4824]: I1124 13:37:24.818655 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": EOF" Nov 24 13:37:25 crc kubenswrapper[4824]: I1124 13:37:25.655201 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerID="e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6" exitCode=143 Nov 24 13:37:25 crc kubenswrapper[4824]: I1124 13:37:25.655281 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f","Type":"ContainerDied","Data":"e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6"} Nov 24 13:37:25 crc kubenswrapper[4824]: I1124 13:37:25.657238 4824 generic.go:334] "Generic (PLEG): container finished" podID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerID="ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7" exitCode=143 Nov 24 13:37:25 crc kubenswrapper[4824]: I1124 13:37:25.657306 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7972c42a-2af1-4d7c-81a6-cdc342978f8b","Type":"ContainerDied","Data":"ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7"} Nov 24 13:37:25 crc kubenswrapper[4824]: I1124 13:37:25.658937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hgdqw" event={"ID":"a2a19c34-3ca4-478e-a544-d9a5e48abb42","Type":"ContainerStarted","Data":"f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b"} Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.419900 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.536645 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nkkd\" (UniqueName: \"kubernetes.io/projected/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-kube-api-access-4nkkd\") pod \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.536723 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-config-data\") pod \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.536769 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-combined-ca-bundle\") pod \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\" (UID: \"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b\") " Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.565033 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-kube-api-access-4nkkd" (OuterVolumeSpecName: "kube-api-access-4nkkd") pod "9661f7ee-75ed-4c09-be2d-d08d2c8eda1b" (UID: "9661f7ee-75ed-4c09-be2d-d08d2c8eda1b"). InnerVolumeSpecName "kube-api-access-4nkkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.577995 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9661f7ee-75ed-4c09-be2d-d08d2c8eda1b" (UID: "9661f7ee-75ed-4c09-be2d-d08d2c8eda1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.583615 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-config-data" (OuterVolumeSpecName: "config-data") pod "9661f7ee-75ed-4c09-be2d-d08d2c8eda1b" (UID: "9661f7ee-75ed-4c09-be2d-d08d2c8eda1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.638615 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nkkd\" (UniqueName: \"kubernetes.io/projected/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-kube-api-access-4nkkd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.638648 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.638657 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.667563 4824 generic.go:334] "Generic (PLEG): container finished" podID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerID="f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b" exitCode=0 Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.667624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hgdqw" event={"ID":"a2a19c34-3ca4-478e-a544-d9a5e48abb42","Type":"ContainerDied","Data":"f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b"} Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.676028 4824 generic.go:334] "Generic (PLEG): container finished" podID="9661f7ee-75ed-4c09-be2d-d08d2c8eda1b" containerID="fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39" exitCode=0 Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.676080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b","Type":"ContainerDied","Data":"fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39"} Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.676109 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9661f7ee-75ed-4c09-be2d-d08d2c8eda1b","Type":"ContainerDied","Data":"a68568fa73f85e68753822e99baa3d533c77a0dc0075d98a227a5885cda1c246"} Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.676125 4824 scope.go:117] "RemoveContainer" containerID="fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.676163 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.729976 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.749538 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.768369 4824 scope.go:117] "RemoveContainer" containerID="fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39" Nov 24 13:37:26 crc kubenswrapper[4824]: E1124 13:37:26.768928 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39\": container with ID starting with fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39 not found: ID does not exist" containerID="fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.768957 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39"} err="failed to get container status \"fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39\": rpc error: code = NotFound desc = could not find container \"fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39\": container with ID starting with fc8403ff0eb4bb8460ef2b9659a8430d3ac93ffd83eb8bce43651bd12340cd39 not found: ID does not exist" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.774856 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:37:26 crc kubenswrapper[4824]: E1124 13:37:26.775272 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf" containerName="nova-manage" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.775288 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf" containerName="nova-manage" Nov 24 13:37:26 crc kubenswrapper[4824]: E1124 13:37:26.775304 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9661f7ee-75ed-4c09-be2d-d08d2c8eda1b" containerName="nova-scheduler-scheduler" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.775311 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9661f7ee-75ed-4c09-be2d-d08d2c8eda1b" containerName="nova-scheduler-scheduler" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.775499 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf" containerName="nova-manage" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.775526 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9661f7ee-75ed-4c09-be2d-d08d2c8eda1b" containerName="nova-scheduler-scheduler" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.776347 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.780953 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.786704 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.843171 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c22644-2997-44ff-ab0d-13ce1a067aa2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"51c22644-2997-44ff-ab0d-13ce1a067aa2\") " pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.843494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c22644-2997-44ff-ab0d-13ce1a067aa2-config-data\") pod \"nova-scheduler-0\" (UID: \"51c22644-2997-44ff-ab0d-13ce1a067aa2\") " pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.843595 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xs77\" (UniqueName: \"kubernetes.io/projected/51c22644-2997-44ff-ab0d-13ce1a067aa2-kube-api-access-6xs77\") pod \"nova-scheduler-0\" (UID: \"51c22644-2997-44ff-ab0d-13ce1a067aa2\") " pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.952309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xs77\" (UniqueName: \"kubernetes.io/projected/51c22644-2997-44ff-ab0d-13ce1a067aa2-kube-api-access-6xs77\") pod \"nova-scheduler-0\" (UID: \"51c22644-2997-44ff-ab0d-13ce1a067aa2\") " pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.952449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c22644-2997-44ff-ab0d-13ce1a067aa2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"51c22644-2997-44ff-ab0d-13ce1a067aa2\") " pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.952477 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c22644-2997-44ff-ab0d-13ce1a067aa2-config-data\") pod \"nova-scheduler-0\" (UID: \"51c22644-2997-44ff-ab0d-13ce1a067aa2\") " pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.960190 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c22644-2997-44ff-ab0d-13ce1a067aa2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"51c22644-2997-44ff-ab0d-13ce1a067aa2\") " pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.961050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c22644-2997-44ff-ab0d-13ce1a067aa2-config-data\") pod \"nova-scheduler-0\" (UID: \"51c22644-2997-44ff-ab0d-13ce1a067aa2\") " pod="openstack/nova-scheduler-0" Nov 24 13:37:26 crc kubenswrapper[4824]: I1124 13:37:26.974968 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xs77\" (UniqueName: \"kubernetes.io/projected/51c22644-2997-44ff-ab0d-13ce1a067aa2-kube-api-access-6xs77\") pod \"nova-scheduler-0\" (UID: \"51c22644-2997-44ff-ab0d-13ce1a067aa2\") " pod="openstack/nova-scheduler-0" Nov 24 13:37:27 crc kubenswrapper[4824]: I1124 13:37:27.025424 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9661f7ee-75ed-4c09-be2d-d08d2c8eda1b" path="/var/lib/kubelet/pods/9661f7ee-75ed-4c09-be2d-d08d2c8eda1b/volumes" Nov 24 13:37:27 crc kubenswrapper[4824]: I1124 13:37:27.096432 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 13:37:27 crc kubenswrapper[4824]: I1124 13:37:27.559514 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 13:37:27 crc kubenswrapper[4824]: I1124 13:37:27.685717 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"51c22644-2997-44ff-ab0d-13ce1a067aa2","Type":"ContainerStarted","Data":"f63f5f742754f58a53d3c9d0fbec2e9cb45e2028c67707cf5c7527062ee1d50f"} Nov 24 13:37:27 crc kubenswrapper[4824]: I1124 13:37:27.690851 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hgdqw" event={"ID":"a2a19c34-3ca4-478e-a544-d9a5e48abb42","Type":"ContainerStarted","Data":"994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647"} Nov 24 13:37:27 crc kubenswrapper[4824]: I1124 13:37:27.713568 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hgdqw" podStartSLOduration=1.917955719 podStartE2EDuration="4.713548715s" podCreationTimestamp="2025-11-24 13:37:23 +0000 UTC" firstStartedPulling="2025-11-24 13:37:24.614389246 +0000 UTC m=+1386.253928556" lastFinishedPulling="2025-11-24 13:37:27.409982242 +0000 UTC m=+1389.049521552" observedRunningTime="2025-11-24 13:37:27.707105334 +0000 UTC m=+1389.346644644" watchObservedRunningTime="2025-11-24 13:37:27.713548715 +0000 UTC m=+1389.353088035" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.383572 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.496773 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-nova-metadata-tls-certs\") pod \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.497158 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsfzr\" (UniqueName: \"kubernetes.io/projected/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-kube-api-access-wsfzr\") pod \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.497200 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-config-data\") pod \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.497231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-logs\") pod \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.497274 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-combined-ca-bundle\") pod \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\" (UID: \"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f\") " Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.499546 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-logs" (OuterVolumeSpecName: "logs") pod "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" (UID: "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.502798 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-kube-api-access-wsfzr" (OuterVolumeSpecName: "kube-api-access-wsfzr") pod "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" (UID: "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f"). InnerVolumeSpecName "kube-api-access-wsfzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.535390 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-config-data" (OuterVolumeSpecName: "config-data") pod "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" (UID: "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.559058 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" (UID: "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.572486 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" (UID: "fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.599098 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.599136 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsfzr\" (UniqueName: \"kubernetes.io/projected/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-kube-api-access-wsfzr\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.599146 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.599156 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.599167 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.702588 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerID="52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75" exitCode=0 Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.702668 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f","Type":"ContainerDied","Data":"52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75"} Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.702707 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f","Type":"ContainerDied","Data":"6451790550c7ecd6971d2fc31ccd7fdde3a67a2fa94f1e4482149e222559eec1"} Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.702728 4824 scope.go:117] "RemoveContainer" containerID="52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.702671 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.704987 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"51c22644-2997-44ff-ab0d-13ce1a067aa2","Type":"ContainerStarted","Data":"ffb2d2a0e19a9f8958892ded55cf0214aa3a4719b4c1b86e762c71b5203790c2"} Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.743793 4824 scope.go:117] "RemoveContainer" containerID="e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.766360 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.766338658 podStartE2EDuration="2.766338658s" podCreationTimestamp="2025-11-24 13:37:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:37:28.737916459 +0000 UTC m=+1390.377455769" watchObservedRunningTime="2025-11-24 13:37:28.766338658 +0000 UTC m=+1390.405877978" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.795293 4824 scope.go:117] "RemoveContainer" containerID="52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75" Nov 24 13:37:28 crc kubenswrapper[4824]: E1124 13:37:28.801026 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75\": container with ID starting with 52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75 not found: ID does not exist" containerID="52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.801240 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75"} err="failed to get container status \"52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75\": rpc error: code = NotFound desc = could not find container \"52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75\": container with ID starting with 52d073cb87d3ce372c3ae36e049bd65a52fceb7d1cd0ccdd65f9945ff925cb75 not found: ID does not exist" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.801365 4824 scope.go:117] "RemoveContainer" containerID="e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.802958 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:28 crc kubenswrapper[4824]: E1124 13:37:28.807002 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6\": container with ID starting with e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6 not found: ID does not exist" containerID="e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.807053 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6"} err="failed to get container status \"e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6\": rpc error: code = NotFound desc = could not find container \"e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6\": container with ID starting with e9a0b65df86ce9d0531ba261041091a09fa76ada6ba77e04aefd85c1c9bffea6 not found: ID does not exist" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.838948 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.844340 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:28 crc kubenswrapper[4824]: E1124 13:37:28.845116 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-log" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.845138 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-log" Nov 24 13:37:28 crc kubenswrapper[4824]: E1124 13:37:28.845174 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-metadata" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.845181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-metadata" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.845354 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-metadata" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.845377 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" containerName="nova-metadata-log" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.846419 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.852623 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.863090 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.863967 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.907951 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cd73a38-f125-4e8d-b902-f8bc08c1122a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.908010 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cd73a38-f125-4e8d-b902-f8bc08c1122a-logs\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.908030 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cd73a38-f125-4e8d-b902-f8bc08c1122a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.908053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cd73a38-f125-4e8d-b902-f8bc08c1122a-config-data\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:28 crc kubenswrapper[4824]: I1124 13:37:28.908073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4pg7\" (UniqueName: \"kubernetes.io/projected/8cd73a38-f125-4e8d-b902-f8bc08c1122a-kube-api-access-s4pg7\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.009353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cd73a38-f125-4e8d-b902-f8bc08c1122a-logs\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.009400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cd73a38-f125-4e8d-b902-f8bc08c1122a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.009440 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cd73a38-f125-4e8d-b902-f8bc08c1122a-config-data\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.009470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4pg7\" (UniqueName: \"kubernetes.io/projected/8cd73a38-f125-4e8d-b902-f8bc08c1122a-kube-api-access-s4pg7\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.009821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cd73a38-f125-4e8d-b902-f8bc08c1122a-logs\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.010045 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cd73a38-f125-4e8d-b902-f8bc08c1122a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.013414 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cd73a38-f125-4e8d-b902-f8bc08c1122a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.013541 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cd73a38-f125-4e8d-b902-f8bc08c1122a-config-data\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.015867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cd73a38-f125-4e8d-b902-f8bc08c1122a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.020447 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f" path="/var/lib/kubelet/pods/fc6a6b8f-d90e-4546-be92-3bd14fb9bc4f/volumes" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.035318 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4pg7\" (UniqueName: \"kubernetes.io/projected/8cd73a38-f125-4e8d-b902-f8bc08c1122a-kube-api-access-s4pg7\") pod \"nova-metadata-0\" (UID: \"8cd73a38-f125-4e8d-b902-f8bc08c1122a\") " pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.191951 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 13:37:29 crc kubenswrapper[4824]: W1124 13:37:29.703472 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cd73a38_f125_4e8d_b902_f8bc08c1122a.slice/crio-07d93f0b09bfb90bbc54f93a05d59e5cb58870ae83b34fdb1a71a7bbacf2bd40 WatchSource:0}: Error finding container 07d93f0b09bfb90bbc54f93a05d59e5cb58870ae83b34fdb1a71a7bbacf2bd40: Status 404 returned error can't find the container with id 07d93f0b09bfb90bbc54f93a05d59e5cb58870ae83b34fdb1a71a7bbacf2bd40 Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.716996 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 13:37:29 crc kubenswrapper[4824]: I1124 13:37:29.729785 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8cd73a38-f125-4e8d-b902-f8bc08c1122a","Type":"ContainerStarted","Data":"07d93f0b09bfb90bbc54f93a05d59e5cb58870ae83b34fdb1a71a7bbacf2bd40"} Nov 24 13:37:30 crc kubenswrapper[4824]: I1124 13:37:30.741832 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8cd73a38-f125-4e8d-b902-f8bc08c1122a","Type":"ContainerStarted","Data":"7b9a34168db7592efabe99727ac92573739d530101ceca9ba20142d5050287de"} Nov 24 13:37:30 crc kubenswrapper[4824]: I1124 13:37:30.742166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8cd73a38-f125-4e8d-b902-f8bc08c1122a","Type":"ContainerStarted","Data":"aa22f6e468b4c81798ef7e67fb3aa222e9924916150b523eeb607356c2a685db"} Nov 24 13:37:30 crc kubenswrapper[4824]: I1124 13:37:30.771991 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.771965635 podStartE2EDuration="2.771965635s" podCreationTimestamp="2025-11-24 13:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:37:30.756732843 +0000 UTC m=+1392.396272153" watchObservedRunningTime="2025-11-24 13:37:30.771965635 +0000 UTC m=+1392.411504975" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.726669 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.760146 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7972c42a-2af1-4d7c-81a6-cdc342978f8b-logs\") pod \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.760243 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rllzq\" (UniqueName: \"kubernetes.io/projected/7972c42a-2af1-4d7c-81a6-cdc342978f8b-kube-api-access-rllzq\") pod \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.760385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-config-data\") pod \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.760418 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-internal-tls-certs\") pod \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.760442 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-combined-ca-bundle\") pod \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.760474 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-public-tls-certs\") pod \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\" (UID: \"7972c42a-2af1-4d7c-81a6-cdc342978f8b\") " Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.771718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7972c42a-2af1-4d7c-81a6-cdc342978f8b-logs" (OuterVolumeSpecName: "logs") pod "7972c42a-2af1-4d7c-81a6-cdc342978f8b" (UID: "7972c42a-2af1-4d7c-81a6-cdc342978f8b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.776861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7972c42a-2af1-4d7c-81a6-cdc342978f8b-kube-api-access-rllzq" (OuterVolumeSpecName: "kube-api-access-rllzq") pod "7972c42a-2af1-4d7c-81a6-cdc342978f8b" (UID: "7972c42a-2af1-4d7c-81a6-cdc342978f8b"). InnerVolumeSpecName "kube-api-access-rllzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.777355 4824 generic.go:334] "Generic (PLEG): container finished" podID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerID="b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e" exitCode=0 Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.777529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7972c42a-2af1-4d7c-81a6-cdc342978f8b","Type":"ContainerDied","Data":"b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e"} Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.777583 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7972c42a-2af1-4d7c-81a6-cdc342978f8b","Type":"ContainerDied","Data":"1be3a431d1a61341162d7332e376e0c853cd9d4bbc0a2f00fce98e5ba2edc9d5"} Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.777594 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.777603 4824 scope.go:117] "RemoveContainer" containerID="b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.836635 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-config-data" (OuterVolumeSpecName: "config-data") pod "7972c42a-2af1-4d7c-81a6-cdc342978f8b" (UID: "7972c42a-2af1-4d7c-81a6-cdc342978f8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.849505 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7972c42a-2af1-4d7c-81a6-cdc342978f8b" (UID: "7972c42a-2af1-4d7c-81a6-cdc342978f8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.856970 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7972c42a-2af1-4d7c-81a6-cdc342978f8b" (UID: "7972c42a-2af1-4d7c-81a6-cdc342978f8b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.862036 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7972c42a-2af1-4d7c-81a6-cdc342978f8b" (UID: "7972c42a-2af1-4d7c-81a6-cdc342978f8b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.862914 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rllzq\" (UniqueName: \"kubernetes.io/projected/7972c42a-2af1-4d7c-81a6-cdc342978f8b-kube-api-access-rllzq\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.862936 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.862945 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.862954 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.862966 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7972c42a-2af1-4d7c-81a6-cdc342978f8b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.862975 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7972c42a-2af1-4d7c-81a6-cdc342978f8b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.889260 4824 scope.go:117] "RemoveContainer" containerID="ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.910480 4824 scope.go:117] "RemoveContainer" containerID="b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e" Nov 24 13:37:31 crc kubenswrapper[4824]: E1124 13:37:31.910903 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e\": container with ID starting with b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e not found: ID does not exist" containerID="b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.910935 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e"} err="failed to get container status \"b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e\": rpc error: code = NotFound desc = could not find container \"b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e\": container with ID starting with b33e1dacfdc9118d674e8ebefc92082561d97071391697986abe331ebd4bdc2e not found: ID does not exist" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.910955 4824 scope.go:117] "RemoveContainer" containerID="ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7" Nov 24 13:37:31 crc kubenswrapper[4824]: E1124 13:37:31.911205 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7\": container with ID starting with ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7 not found: ID does not exist" containerID="ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7" Nov 24 13:37:31 crc kubenswrapper[4824]: I1124 13:37:31.911225 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7"} err="failed to get container status \"ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7\": rpc error: code = NotFound desc = could not find container \"ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7\": container with ID starting with ab6925f39b457811ddc2e6d0d1248ed7bd9ea445cfc67412f336d7b5e364dcb7 not found: ID does not exist" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.096872 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.117538 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.128447 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.155869 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:32 crc kubenswrapper[4824]: E1124 13:37:32.156617 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-log" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.156732 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-log" Nov 24 13:37:32 crc kubenswrapper[4824]: E1124 13:37:32.156874 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-api" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.157116 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-api" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.157434 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-log" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.157561 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" containerName="nova-api-api" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.159294 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.162195 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.162494 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.167425 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.202115 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.270059 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebf76616-c332-4ed1-a188-99843d4692d3-logs\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.270150 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-public-tls-certs\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.270442 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.270504 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-config-data\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.270655 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.270766 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wgkf\" (UniqueName: \"kubernetes.io/projected/ebf76616-c332-4ed1-a188-99843d4692d3-kube-api-access-4wgkf\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.372096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.372171 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wgkf\" (UniqueName: \"kubernetes.io/projected/ebf76616-c332-4ed1-a188-99843d4692d3-kube-api-access-4wgkf\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.372217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebf76616-c332-4ed1-a188-99843d4692d3-logs\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.372302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-public-tls-certs\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.372398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.372427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-config-data\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.374571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebf76616-c332-4ed1-a188-99843d4692d3-logs\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.379010 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-config-data\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.379172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.379943 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-public-tls-certs\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.379988 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebf76616-c332-4ed1-a188-99843d4692d3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.398943 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wgkf\" (UniqueName: \"kubernetes.io/projected/ebf76616-c332-4ed1-a188-99843d4692d3-kube-api-access-4wgkf\") pod \"nova-api-0\" (UID: \"ebf76616-c332-4ed1-a188-99843d4692d3\") " pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.483751 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 13:37:32 crc kubenswrapper[4824]: I1124 13:37:32.982432 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 13:37:32 crc kubenswrapper[4824]: W1124 13:37:32.983326 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebf76616_c332_4ed1_a188_99843d4692d3.slice/crio-d14594490744d4e42353212fbd824888cce1aebae9b15a5230f9d84184a73824 WatchSource:0}: Error finding container d14594490744d4e42353212fbd824888cce1aebae9b15a5230f9d84184a73824: Status 404 returned error can't find the container with id d14594490744d4e42353212fbd824888cce1aebae9b15a5230f9d84184a73824 Nov 24 13:37:33 crc kubenswrapper[4824]: I1124 13:37:33.040690 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7972c42a-2af1-4d7c-81a6-cdc342978f8b" path="/var/lib/kubelet/pods/7972c42a-2af1-4d7c-81a6-cdc342978f8b/volumes" Nov 24 13:37:33 crc kubenswrapper[4824]: I1124 13:37:33.730627 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:33 crc kubenswrapper[4824]: I1124 13:37:33.731349 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:33 crc kubenswrapper[4824]: I1124 13:37:33.785943 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:33 crc kubenswrapper[4824]: I1124 13:37:33.829184 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ebf76616-c332-4ed1-a188-99843d4692d3","Type":"ContainerStarted","Data":"8406465db0596388d17544c7de39242af06be4c60d52363bbe27ec04e558f7c1"} Nov 24 13:37:33 crc kubenswrapper[4824]: I1124 13:37:33.829474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ebf76616-c332-4ed1-a188-99843d4692d3","Type":"ContainerStarted","Data":"e65456c7424f75ecf28088572a45a65e361de1a86fba80b339f18cb35e343e69"} Nov 24 13:37:33 crc kubenswrapper[4824]: I1124 13:37:33.829539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ebf76616-c332-4ed1-a188-99843d4692d3","Type":"ContainerStarted","Data":"d14594490744d4e42353212fbd824888cce1aebae9b15a5230f9d84184a73824"} Nov 24 13:37:33 crc kubenswrapper[4824]: I1124 13:37:33.857567 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.8575456080000001 podStartE2EDuration="1.857545608s" podCreationTimestamp="2025-11-24 13:37:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:37:33.849303558 +0000 UTC m=+1395.488842888" watchObservedRunningTime="2025-11-24 13:37:33.857545608 +0000 UTC m=+1395.497084928" Nov 24 13:37:33 crc kubenswrapper[4824]: I1124 13:37:33.886952 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:34 crc kubenswrapper[4824]: I1124 13:37:34.035934 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hgdqw"] Nov 24 13:37:34 crc kubenswrapper[4824]: I1124 13:37:34.192492 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:37:34 crc kubenswrapper[4824]: I1124 13:37:34.192754 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 13:37:35 crc kubenswrapper[4824]: I1124 13:37:35.854567 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hgdqw" podUID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerName="registry-server" containerID="cri-o://994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647" gracePeriod=2 Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.342037 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.446319 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-utilities\") pod \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.446528 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k47hd\" (UniqueName: \"kubernetes.io/projected/a2a19c34-3ca4-478e-a544-d9a5e48abb42-kube-api-access-k47hd\") pod \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.446580 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-catalog-content\") pod \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\" (UID: \"a2a19c34-3ca4-478e-a544-d9a5e48abb42\") " Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.447472 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-utilities" (OuterVolumeSpecName: "utilities") pod "a2a19c34-3ca4-478e-a544-d9a5e48abb42" (UID: "a2a19c34-3ca4-478e-a544-d9a5e48abb42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.452114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2a19c34-3ca4-478e-a544-d9a5e48abb42-kube-api-access-k47hd" (OuterVolumeSpecName: "kube-api-access-k47hd") pod "a2a19c34-3ca4-478e-a544-d9a5e48abb42" (UID: "a2a19c34-3ca4-478e-a544-d9a5e48abb42"). InnerVolumeSpecName "kube-api-access-k47hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.472951 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2a19c34-3ca4-478e-a544-d9a5e48abb42" (UID: "a2a19c34-3ca4-478e-a544-d9a5e48abb42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.548410 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k47hd\" (UniqueName: \"kubernetes.io/projected/a2a19c34-3ca4-478e-a544-d9a5e48abb42-kube-api-access-k47hd\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.548445 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.548454 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2a19c34-3ca4-478e-a544-d9a5e48abb42-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.867880 4824 generic.go:334] "Generic (PLEG): container finished" podID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerID="994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647" exitCode=0 Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.867923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hgdqw" event={"ID":"a2a19c34-3ca4-478e-a544-d9a5e48abb42","Type":"ContainerDied","Data":"994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647"} Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.867930 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hgdqw" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.867960 4824 scope.go:117] "RemoveContainer" containerID="994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.867950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hgdqw" event={"ID":"a2a19c34-3ca4-478e-a544-d9a5e48abb42","Type":"ContainerDied","Data":"ffc77ea27c83ed44df8b54ca8f45a072d415fa85f48295f8eb234dbded782efc"} Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.896348 4824 scope.go:117] "RemoveContainer" containerID="f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.916118 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hgdqw"] Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.924115 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hgdqw"] Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.937396 4824 scope.go:117] "RemoveContainer" containerID="71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.992240 4824 scope.go:117] "RemoveContainer" containerID="994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647" Nov 24 13:37:36 crc kubenswrapper[4824]: E1124 13:37:36.994263 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647\": container with ID starting with 994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647 not found: ID does not exist" containerID="994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.994303 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647"} err="failed to get container status \"994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647\": rpc error: code = NotFound desc = could not find container \"994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647\": container with ID starting with 994b12af19d5fa1ad68b009dc0eaa337110c2d7296720ba04b4fc1eb698e1647 not found: ID does not exist" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.994328 4824 scope.go:117] "RemoveContainer" containerID="f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b" Nov 24 13:37:36 crc kubenswrapper[4824]: E1124 13:37:36.994881 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b\": container with ID starting with f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b not found: ID does not exist" containerID="f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.994967 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b"} err="failed to get container status \"f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b\": rpc error: code = NotFound desc = could not find container \"f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b\": container with ID starting with f004615d9c93aa10504ac18a69367267c5cb09ef324170d2c8db2a915174fd2b not found: ID does not exist" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.995038 4824 scope.go:117] "RemoveContainer" containerID="71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232" Nov 24 13:37:36 crc kubenswrapper[4824]: E1124 13:37:36.995390 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232\": container with ID starting with 71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232 not found: ID does not exist" containerID="71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232" Nov 24 13:37:36 crc kubenswrapper[4824]: I1124 13:37:36.995429 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232"} err="failed to get container status \"71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232\": rpc error: code = NotFound desc = could not find container \"71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232\": container with ID starting with 71bf44c0f8f0379a4d758044729486436e72675380d1f63b8aee6594bc30a232 not found: ID does not exist" Nov 24 13:37:37 crc kubenswrapper[4824]: I1124 13:37:37.030026 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" path="/var/lib/kubelet/pods/a2a19c34-3ca4-478e-a544-d9a5e48abb42/volumes" Nov 24 13:37:37 crc kubenswrapper[4824]: I1124 13:37:37.097469 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 13:37:37 crc kubenswrapper[4824]: I1124 13:37:37.140767 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 13:37:37 crc kubenswrapper[4824]: I1124 13:37:37.936187 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 13:37:39 crc kubenswrapper[4824]: I1124 13:37:39.193419 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 13:37:39 crc kubenswrapper[4824]: I1124 13:37:39.193870 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 13:37:40 crc kubenswrapper[4824]: I1124 13:37:40.213026 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8cd73a38-f125-4e8d-b902-f8bc08c1122a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:37:40 crc kubenswrapper[4824]: I1124 13:37:40.213024 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8cd73a38-f125-4e8d-b902-f8bc08c1122a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:37:42 crc kubenswrapper[4824]: I1124 13:37:42.485499 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:37:42 crc kubenswrapper[4824]: I1124 13:37:42.485859 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 13:37:43 crc kubenswrapper[4824]: I1124 13:37:43.501040 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ebf76616-c332-4ed1-a188-99843d4692d3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:37:43 crc kubenswrapper[4824]: I1124 13:37:43.501064 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ebf76616-c332-4ed1-a188-99843d4692d3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 13:37:45 crc kubenswrapper[4824]: I1124 13:37:45.968533 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 13:37:49 crc kubenswrapper[4824]: I1124 13:37:49.199071 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 13:37:49 crc kubenswrapper[4824]: I1124 13:37:49.203698 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 13:37:49 crc kubenswrapper[4824]: I1124 13:37:49.212209 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.016055 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.466640 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vbmpp"] Nov 24 13:37:50 crc kubenswrapper[4824]: E1124 13:37:50.467167 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerName="extract-content" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.467185 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerName="extract-content" Nov 24 13:37:50 crc kubenswrapper[4824]: E1124 13:37:50.467208 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerName="extract-utilities" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.467216 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerName="extract-utilities" Nov 24 13:37:50 crc kubenswrapper[4824]: E1124 13:37:50.467230 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerName="registry-server" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.467237 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerName="registry-server" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.467456 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2a19c34-3ca4-478e-a544-d9a5e48abb42" containerName="registry-server" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.469070 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.481111 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vbmpp"] Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.614909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mnp6\" (UniqueName: \"kubernetes.io/projected/a5c0f777-017e-4db2-8947-d27a71546d86-kube-api-access-9mnp6\") pod \"redhat-operators-vbmpp\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.615868 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-catalog-content\") pod \"redhat-operators-vbmpp\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.615939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-utilities\") pod \"redhat-operators-vbmpp\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.717252 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-catalog-content\") pod \"redhat-operators-vbmpp\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.717327 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-utilities\") pod \"redhat-operators-vbmpp\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.717382 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mnp6\" (UniqueName: \"kubernetes.io/projected/a5c0f777-017e-4db2-8947-d27a71546d86-kube-api-access-9mnp6\") pod \"redhat-operators-vbmpp\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.717850 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-utilities\") pod \"redhat-operators-vbmpp\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.719135 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-catalog-content\") pod \"redhat-operators-vbmpp\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.747447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mnp6\" (UniqueName: \"kubernetes.io/projected/a5c0f777-017e-4db2-8947-d27a71546d86-kube-api-access-9mnp6\") pod \"redhat-operators-vbmpp\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:50 crc kubenswrapper[4824]: I1124 13:37:50.800702 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:37:51 crc kubenswrapper[4824]: W1124 13:37:51.794533 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5c0f777_017e_4db2_8947_d27a71546d86.slice/crio-ebd232fbf156b5fa1a5b285abc0a1dd321e5c2f8931ed3aad59ec4dd8f1ace06 WatchSource:0}: Error finding container ebd232fbf156b5fa1a5b285abc0a1dd321e5c2f8931ed3aad59ec4dd8f1ace06: Status 404 returned error can't find the container with id ebd232fbf156b5fa1a5b285abc0a1dd321e5c2f8931ed3aad59ec4dd8f1ace06 Nov 24 13:37:51 crc kubenswrapper[4824]: I1124 13:37:51.802383 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vbmpp"] Nov 24 13:37:52 crc kubenswrapper[4824]: I1124 13:37:52.064883 4824 generic.go:334] "Generic (PLEG): container finished" podID="a5c0f777-017e-4db2-8947-d27a71546d86" containerID="6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9" exitCode=0 Nov 24 13:37:52 crc kubenswrapper[4824]: I1124 13:37:52.066743 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbmpp" event={"ID":"a5c0f777-017e-4db2-8947-d27a71546d86","Type":"ContainerDied","Data":"6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9"} Nov 24 13:37:52 crc kubenswrapper[4824]: I1124 13:37:52.066879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbmpp" event={"ID":"a5c0f777-017e-4db2-8947-d27a71546d86","Type":"ContainerStarted","Data":"ebd232fbf156b5fa1a5b285abc0a1dd321e5c2f8931ed3aad59ec4dd8f1ace06"} Nov 24 13:37:52 crc kubenswrapper[4824]: I1124 13:37:52.494337 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 13:37:52 crc kubenswrapper[4824]: I1124 13:37:52.495163 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 13:37:52 crc kubenswrapper[4824]: I1124 13:37:52.506221 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 13:37:52 crc kubenswrapper[4824]: I1124 13:37:52.510373 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 13:37:53 crc kubenswrapper[4824]: I1124 13:37:53.080442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbmpp" event={"ID":"a5c0f777-017e-4db2-8947-d27a71546d86","Type":"ContainerStarted","Data":"9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363"} Nov 24 13:37:53 crc kubenswrapper[4824]: I1124 13:37:53.081044 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 13:37:53 crc kubenswrapper[4824]: I1124 13:37:53.087192 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 13:37:58 crc kubenswrapper[4824]: I1124 13:37:58.128580 4824 generic.go:334] "Generic (PLEG): container finished" podID="a5c0f777-017e-4db2-8947-d27a71546d86" containerID="9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363" exitCode=0 Nov 24 13:37:58 crc kubenswrapper[4824]: I1124 13:37:58.128656 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbmpp" event={"ID":"a5c0f777-017e-4db2-8947-d27a71546d86","Type":"ContainerDied","Data":"9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363"} Nov 24 13:37:59 crc kubenswrapper[4824]: I1124 13:37:59.139361 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbmpp" event={"ID":"a5c0f777-017e-4db2-8947-d27a71546d86","Type":"ContainerStarted","Data":"e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e"} Nov 24 13:37:59 crc kubenswrapper[4824]: I1124 13:37:59.163207 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vbmpp" podStartSLOduration=2.6977917810000003 podStartE2EDuration="9.163191325s" podCreationTimestamp="2025-11-24 13:37:50 +0000 UTC" firstStartedPulling="2025-11-24 13:37:52.066495169 +0000 UTC m=+1413.706034499" lastFinishedPulling="2025-11-24 13:37:58.531894733 +0000 UTC m=+1420.171434043" observedRunningTime="2025-11-24 13:37:59.162318646 +0000 UTC m=+1420.801857956" watchObservedRunningTime="2025-11-24 13:37:59.163191325 +0000 UTC m=+1420.802730635" Nov 24 13:38:00 crc kubenswrapper[4824]: I1124 13:38:00.801846 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:38:00 crc kubenswrapper[4824]: I1124 13:38:00.802238 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:38:01 crc kubenswrapper[4824]: I1124 13:38:01.791364 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:38:01 crc kubenswrapper[4824]: I1124 13:38:01.851925 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vbmpp" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="registry-server" probeResult="failure" output=< Nov 24 13:38:01 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 13:38:01 crc kubenswrapper[4824]: > Nov 24 13:38:02 crc kubenswrapper[4824]: I1124 13:38:02.680489 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:38:06 crc kubenswrapper[4824]: I1124 13:38:06.998004 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="bc596e02-b760-49d3-a0c2-00ec8e751e05" containerName="rabbitmq" containerID="cri-o://f95a2725a5975c99b2ebd62bef582367f706a83ba95e987fe32fd34e07d0c268" gracePeriod=604796 Nov 24 13:38:07 crc kubenswrapper[4824]: I1124 13:38:07.067651 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" containerName="rabbitmq" containerID="cri-o://c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126" gracePeriod=604795 Nov 24 13:38:08 crc kubenswrapper[4824]: I1124 13:38:08.108634 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Nov 24 13:38:08 crc kubenswrapper[4824]: I1124 13:38:08.722540 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="bc596e02-b760-49d3-a0c2-00ec8e751e05" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Nov 24 13:38:10 crc kubenswrapper[4824]: I1124 13:38:10.788010 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:38:10 crc kubenswrapper[4824]: I1124 13:38:10.788367 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:38:11 crc kubenswrapper[4824]: I1124 13:38:11.843421 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vbmpp" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="registry-server" probeResult="failure" output=< Nov 24 13:38:11 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 13:38:11 crc kubenswrapper[4824]: > Nov 24 13:38:13 crc kubenswrapper[4824]: I1124 13:38:13.282674 4824 generic.go:334] "Generic (PLEG): container finished" podID="bc596e02-b760-49d3-a0c2-00ec8e751e05" containerID="f95a2725a5975c99b2ebd62bef582367f706a83ba95e987fe32fd34e07d0c268" exitCode=0 Nov 24 13:38:13 crc kubenswrapper[4824]: I1124 13:38:13.282728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bc596e02-b760-49d3-a0c2-00ec8e751e05","Type":"ContainerDied","Data":"f95a2725a5975c99b2ebd62bef582367f706a83ba95e987fe32fd34e07d0c268"} Nov 24 13:38:13 crc kubenswrapper[4824]: I1124 13:38:13.842859 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:13 crc kubenswrapper[4824]: I1124 13:38:13.856053 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006113 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-plugins-conf\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006158 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-tls\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006189 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-plugins-conf\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006212 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-erlang-cookie\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-config-data\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006335 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb4jb\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-kube-api-access-jb4jb\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006360 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-confd\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006403 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-plugins\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-plugins\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006446 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc596e02-b760-49d3-a0c2-00ec8e751e05-erlang-cookie-secret\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006465 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-erlang-cookie-secret\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006485 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-tls\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006531 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-confd\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006547 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-server-conf\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006596 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc596e02-b760-49d3-a0c2-00ec8e751e05-pod-info\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006603 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006642 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-erlang-cookie\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006611 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006691 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-pod-info\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006709 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-server-conf\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006765 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ds8w\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-kube-api-access-2ds8w\") pod \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\" (UID: \"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-config-data\") pod \"bc596e02-b760-49d3-a0c2-00ec8e751e05\" (UID: \"bc596e02-b760-49d3-a0c2-00ec8e751e05\") " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.006952 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.007199 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.007216 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.007219 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.007227 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.008225 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.008608 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.020908 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/bc596e02-b760-49d3-a0c2-00ec8e751e05-pod-info" (OuterVolumeSpecName: "pod-info") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.020970 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-pod-info" (OuterVolumeSpecName: "pod-info") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.025121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.031413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.032246 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.034084 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc596e02-b760-49d3-a0c2-00ec8e751e05-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.051447 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.052246 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.052379 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-kube-api-access-2ds8w" (OuterVolumeSpecName: "kube-api-access-2ds8w") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "kube-api-access-2ds8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.072144 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-kube-api-access-jb4jb" (OuterVolumeSpecName: "kube-api-access-jb4jb") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "kube-api-access-jb4jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.077516 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-config-data" (OuterVolumeSpecName: "config-data") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.090296 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-config-data" (OuterVolumeSpecName: "config-data") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112014 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc596e02-b760-49d3-a0c2-00ec8e751e05-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112056 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112068 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112079 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc596e02-b760-49d3-a0c2-00ec8e751e05-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112090 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112129 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112141 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112158 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112178 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ds8w\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-kube-api-access-2ds8w\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112187 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112197 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112208 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112218 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112228 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb4jb\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-kube-api-access-jb4jb\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.112239 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.138433 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.161946 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.169332 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-server-conf" (OuterVolumeSpecName: "server-conf") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.215314 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.215342 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc596e02-b760-49d3-a0c2-00ec8e751e05-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.215352 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.221181 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-server-conf" (OuterVolumeSpecName: "server-conf") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.252887 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" (UID: "9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.278921 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "bc596e02-b760-49d3-a0c2-00ec8e751e05" (UID: "bc596e02-b760-49d3-a0c2-00ec8e751e05"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.302878 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bc596e02-b760-49d3-a0c2-00ec8e751e05","Type":"ContainerDied","Data":"2982de524dd519d04b87f72eb9b3808a8c6550b9a6b1a5bdb14d0f2158f18dd8"} Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.302939 4824 scope.go:117] "RemoveContainer" containerID="f95a2725a5975c99b2ebd62bef582367f706a83ba95e987fe32fd34e07d0c268" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.303186 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.315308 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" containerID="c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126" exitCode=0 Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.315345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7","Type":"ContainerDied","Data":"c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126"} Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.315370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7","Type":"ContainerDied","Data":"dfeb12863341e17eefcdd077ee34f2e1a72a4defa9795a5511355e904d743376"} Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.315426 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.318328 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc596e02-b760-49d3-a0c2-00ec8e751e05-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.318352 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.318360 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.381110 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.389083 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.405302 4824 scope.go:117] "RemoveContainer" containerID="46d8ac6888d06952ac6295da8bf804f80e93e9aa76eb2e3004f7277c99076e12" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.429646 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.464847 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.464954 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:38:14 crc kubenswrapper[4824]: E1124 13:38:14.465939 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc596e02-b760-49d3-a0c2-00ec8e751e05" containerName="setup-container" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.466000 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc596e02-b760-49d3-a0c2-00ec8e751e05" containerName="setup-container" Nov 24 13:38:14 crc kubenswrapper[4824]: E1124 13:38:14.466057 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" containerName="rabbitmq" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.466065 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" containerName="rabbitmq" Nov 24 13:38:14 crc kubenswrapper[4824]: E1124 13:38:14.466079 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" containerName="setup-container" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.466098 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" containerName="setup-container" Nov 24 13:38:14 crc kubenswrapper[4824]: E1124 13:38:14.466112 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc596e02-b760-49d3-a0c2-00ec8e751e05" containerName="rabbitmq" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.466118 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc596e02-b760-49d3-a0c2-00ec8e751e05" containerName="rabbitmq" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.466574 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc596e02-b760-49d3-a0c2-00ec8e751e05" containerName="rabbitmq" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.466654 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" containerName="rabbitmq" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.468675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.472838 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mfcwz" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.472979 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.473527 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.473932 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.473973 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.475043 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.475194 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.478878 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.480008 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.484194 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.486751 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.486978 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.487088 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.487188 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.487289 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4xrjn" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.487394 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.504598 4824 scope.go:117] "RemoveContainer" containerID="c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.511251 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529491 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529520 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98ec6c14-59af-4b38-8e01-89ff157036ca-pod-info\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98ec6c14-59af-4b38-8e01-89ff157036ca-server-conf\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529604 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98ec6c14-59af-4b38-8e01-89ff157036ca-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529756 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98ec6c14-59af-4b38-8e01-89ff157036ca-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98ec6c14-59af-4b38-8e01-89ff157036ca-config-data\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.529993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln859\" (UniqueName: \"kubernetes.io/projected/98ec6c14-59af-4b38-8e01-89ff157036ca-kube-api-access-ln859\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.569561 4824 scope.go:117] "RemoveContainer" containerID="3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.574544 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.624371 4824 scope.go:117] "RemoveContainer" containerID="c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126" Nov 24 13:38:14 crc kubenswrapper[4824]: E1124 13:38:14.625112 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126\": container with ID starting with c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126 not found: ID does not exist" containerID="c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.625144 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126"} err="failed to get container status \"c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126\": rpc error: code = NotFound desc = could not find container \"c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126\": container with ID starting with c132d76b7e4e56c573d18ae288dba75f2f7f7a07464f4b7208c38f4b71c23126 not found: ID does not exist" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.625164 4824 scope.go:117] "RemoveContainer" containerID="3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c" Nov 24 13:38:14 crc kubenswrapper[4824]: E1124 13:38:14.625472 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c\": container with ID starting with 3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c not found: ID does not exist" containerID="3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.625492 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c"} err="failed to get container status \"3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c\": rpc error: code = NotFound desc = could not find container \"3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c\": container with ID starting with 3fe6940de52b5c38fbe36d0245dfb4990662dca798033ce579b2314745e4d51c not found: ID does not exist" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632547 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98ec6c14-59af-4b38-8e01-89ff157036ca-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632607 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632634 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/829719ca-f478-456b-9bbc-07cad8469731-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632689 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632729 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/829719ca-f478-456b-9bbc-07cad8469731-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632762 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/829719ca-f478-456b-9bbc-07cad8469731-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632787 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98ec6c14-59af-4b38-8e01-89ff157036ca-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/829719ca-f478-456b-9bbc-07cad8469731-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632881 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98ec6c14-59af-4b38-8e01-89ff157036ca-config-data\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632938 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.632961 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t45p\" (UniqueName: \"kubernetes.io/projected/829719ca-f478-456b-9bbc-07cad8469731-kube-api-access-5t45p\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln859\" (UniqueName: \"kubernetes.io/projected/98ec6c14-59af-4b38-8e01-89ff157036ca-kube-api-access-ln859\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/829719ca-f478-456b-9bbc-07cad8469731-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633085 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633108 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633133 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98ec6c14-59af-4b38-8e01-89ff157036ca-pod-info\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633200 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98ec6c14-59af-4b38-8e01-89ff157036ca-server-conf\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.633786 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.636076 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98ec6c14-59af-4b38-8e01-89ff157036ca-config-data\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.638981 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.639057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.640606 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98ec6c14-59af-4b38-8e01-89ff157036ca-server-conf\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.640883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98ec6c14-59af-4b38-8e01-89ff157036ca-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.644227 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.644329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98ec6c14-59af-4b38-8e01-89ff157036ca-pod-info\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.646641 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/98ec6c14-59af-4b38-8e01-89ff157036ca-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.656222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98ec6c14-59af-4b38-8e01-89ff157036ca-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.659226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln859\" (UniqueName: \"kubernetes.io/projected/98ec6c14-59af-4b38-8e01-89ff157036ca-kube-api-access-ln859\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.685636 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"98ec6c14-59af-4b38-8e01-89ff157036ca\") " pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734626 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/829719ca-f478-456b-9bbc-07cad8469731-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734679 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/829719ca-f478-456b-9bbc-07cad8469731-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/829719ca-f478-456b-9bbc-07cad8469731-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/829719ca-f478-456b-9bbc-07cad8469731-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734790 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t45p\" (UniqueName: \"kubernetes.io/projected/829719ca-f478-456b-9bbc-07cad8469731-kube-api-access-5t45p\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734827 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/829719ca-f478-456b-9bbc-07cad8469731-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.734866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.735193 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.735497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.736356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/829719ca-f478-456b-9bbc-07cad8469731-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.738228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/829719ca-f478-456b-9bbc-07cad8469731-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.738502 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/829719ca-f478-456b-9bbc-07cad8469731-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.739298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.741077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.741304 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/829719ca-f478-456b-9bbc-07cad8469731-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.741421 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/829719ca-f478-456b-9bbc-07cad8469731-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.741671 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/829719ca-f478-456b-9bbc-07cad8469731-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.757685 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t45p\" (UniqueName: \"kubernetes.io/projected/829719ca-f478-456b-9bbc-07cad8469731-kube-api-access-5t45p\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.781965 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"829719ca-f478-456b-9bbc-07cad8469731\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.842076 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 13:38:14 crc kubenswrapper[4824]: I1124 13:38:14.844205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:15 crc kubenswrapper[4824]: I1124 13:38:15.039645 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7" path="/var/lib/kubelet/pods/9f48dd6f-cca1-42fc-b4bd-bd841b8fb6c7/volumes" Nov 24 13:38:15 crc kubenswrapper[4824]: I1124 13:38:15.043377 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc596e02-b760-49d3-a0c2-00ec8e751e05" path="/var/lib/kubelet/pods/bc596e02-b760-49d3-a0c2-00ec8e751e05/volumes" Nov 24 13:38:15 crc kubenswrapper[4824]: I1124 13:38:15.352466 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 13:38:15 crc kubenswrapper[4824]: I1124 13:38:15.450149 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.333192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"98ec6c14-59af-4b38-8e01-89ff157036ca","Type":"ContainerStarted","Data":"7c1f1eff53d38ddd451ec990ecf5f03039c60ae57e1a04b66b8d09920bdb20af"} Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.334555 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"829719ca-f478-456b-9bbc-07cad8469731","Type":"ContainerStarted","Data":"57bcf205fdebf2f2790ab9896b0d2f7fd46d9f3ffb63dbf36d81a72c5f987817"} Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.685761 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dn5hh"] Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.687299 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.691681 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.696030 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dn5hh"] Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.776681 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2ss9\" (UniqueName: \"kubernetes.io/projected/675c8288-8692-4dd4-a66d-09e5447e349d-kube-api-access-c2ss9\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.776992 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.777115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.777240 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.777350 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.777521 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-config\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.777660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.879371 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-config\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.879437 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.879480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.879497 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2ss9\" (UniqueName: \"kubernetes.io/projected/675c8288-8692-4dd4-a66d-09e5447e349d-kube-api-access-c2ss9\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.879545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.879576 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.879602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.880485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.880606 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-config\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.880955 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.881105 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.881168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.881528 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:16 crc kubenswrapper[4824]: I1124 13:38:16.906730 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2ss9\" (UniqueName: \"kubernetes.io/projected/675c8288-8692-4dd4-a66d-09e5447e349d-kube-api-access-c2ss9\") pod \"dnsmasq-dns-79bd4cc8c9-dn5hh\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:17 crc kubenswrapper[4824]: I1124 13:38:17.009942 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:17 crc kubenswrapper[4824]: I1124 13:38:17.345401 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"829719ca-f478-456b-9bbc-07cad8469731","Type":"ContainerStarted","Data":"7e24d3b66b28d5b61bfb4c84e469b05e234cc60e56ea2280447a2aeda9f13b2a"} Nov 24 13:38:17 crc kubenswrapper[4824]: I1124 13:38:17.348284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"98ec6c14-59af-4b38-8e01-89ff157036ca","Type":"ContainerStarted","Data":"3b3364ac731f06487c715445e1125ca85845a9f9db1d5bf271b2b5a8615efc37"} Nov 24 13:38:17 crc kubenswrapper[4824]: I1124 13:38:17.531105 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dn5hh"] Nov 24 13:38:18 crc kubenswrapper[4824]: I1124 13:38:18.359252 4824 generic.go:334] "Generic (PLEG): container finished" podID="675c8288-8692-4dd4-a66d-09e5447e349d" containerID="c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f" exitCode=0 Nov 24 13:38:18 crc kubenswrapper[4824]: I1124 13:38:18.359587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" event={"ID":"675c8288-8692-4dd4-a66d-09e5447e349d","Type":"ContainerDied","Data":"c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f"} Nov 24 13:38:18 crc kubenswrapper[4824]: I1124 13:38:18.359718 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" event={"ID":"675c8288-8692-4dd4-a66d-09e5447e349d","Type":"ContainerStarted","Data":"e9a5c6740a2276271c7fc65b3dc2b0221b8dedfde9cddfa505713b49db8c628e"} Nov 24 13:38:19 crc kubenswrapper[4824]: I1124 13:38:19.368765 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" event={"ID":"675c8288-8692-4dd4-a66d-09e5447e349d","Type":"ContainerStarted","Data":"432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc"} Nov 24 13:38:19 crc kubenswrapper[4824]: I1124 13:38:19.369996 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:19 crc kubenswrapper[4824]: I1124 13:38:19.388557 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" podStartSLOduration=3.388537513 podStartE2EDuration="3.388537513s" podCreationTimestamp="2025-11-24 13:38:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:38:19.385293871 +0000 UTC m=+1441.024833181" watchObservedRunningTime="2025-11-24 13:38:19.388537513 +0000 UTC m=+1441.028076823" Nov 24 13:38:20 crc kubenswrapper[4824]: I1124 13:38:20.851479 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:38:20 crc kubenswrapper[4824]: I1124 13:38:20.905402 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:38:21 crc kubenswrapper[4824]: I1124 13:38:21.662278 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vbmpp"] Nov 24 13:38:22 crc kubenswrapper[4824]: I1124 13:38:22.394694 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vbmpp" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="registry-server" containerID="cri-o://e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e" gracePeriod=2 Nov 24 13:38:22 crc kubenswrapper[4824]: I1124 13:38:22.918021 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:38:22 crc kubenswrapper[4824]: I1124 13:38:22.999516 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-catalog-content\") pod \"a5c0f777-017e-4db2-8947-d27a71546d86\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " Nov 24 13:38:22 crc kubenswrapper[4824]: I1124 13:38:22.999591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mnp6\" (UniqueName: \"kubernetes.io/projected/a5c0f777-017e-4db2-8947-d27a71546d86-kube-api-access-9mnp6\") pod \"a5c0f777-017e-4db2-8947-d27a71546d86\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " Nov 24 13:38:22 crc kubenswrapper[4824]: I1124 13:38:22.999831 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-utilities\") pod \"a5c0f777-017e-4db2-8947-d27a71546d86\" (UID: \"a5c0f777-017e-4db2-8947-d27a71546d86\") " Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.000407 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-utilities" (OuterVolumeSpecName: "utilities") pod "a5c0f777-017e-4db2-8947-d27a71546d86" (UID: "a5c0f777-017e-4db2-8947-d27a71546d86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.005028 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5c0f777-017e-4db2-8947-d27a71546d86-kube-api-access-9mnp6" (OuterVolumeSpecName: "kube-api-access-9mnp6") pod "a5c0f777-017e-4db2-8947-d27a71546d86" (UID: "a5c0f777-017e-4db2-8947-d27a71546d86"). InnerVolumeSpecName "kube-api-access-9mnp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.097913 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5c0f777-017e-4db2-8947-d27a71546d86" (UID: "a5c0f777-017e-4db2-8947-d27a71546d86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.104281 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.104320 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c0f777-017e-4db2-8947-d27a71546d86-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.104331 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mnp6\" (UniqueName: \"kubernetes.io/projected/a5c0f777-017e-4db2-8947-d27a71546d86-kube-api-access-9mnp6\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.404634 4824 generic.go:334] "Generic (PLEG): container finished" podID="a5c0f777-017e-4db2-8947-d27a71546d86" containerID="e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e" exitCode=0 Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.404689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbmpp" event={"ID":"a5c0f777-017e-4db2-8947-d27a71546d86","Type":"ContainerDied","Data":"e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e"} Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.404720 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbmpp" event={"ID":"a5c0f777-017e-4db2-8947-d27a71546d86","Type":"ContainerDied","Data":"ebd232fbf156b5fa1a5b285abc0a1dd321e5c2f8931ed3aad59ec4dd8f1ace06"} Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.404741 4824 scope.go:117] "RemoveContainer" containerID="e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.404770 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vbmpp" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.444565 4824 scope.go:117] "RemoveContainer" containerID="9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.454097 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vbmpp"] Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.462302 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vbmpp"] Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.475423 4824 scope.go:117] "RemoveContainer" containerID="6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.525677 4824 scope.go:117] "RemoveContainer" containerID="e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e" Nov 24 13:38:23 crc kubenswrapper[4824]: E1124 13:38:23.526624 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e\": container with ID starting with e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e not found: ID does not exist" containerID="e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.526664 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e"} err="failed to get container status \"e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e\": rpc error: code = NotFound desc = could not find container \"e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e\": container with ID starting with e47342a0b8edf0675a8c4c7ca4cac51dd9cf3f08646431e842433c6d4609a57e not found: ID does not exist" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.526691 4824 scope.go:117] "RemoveContainer" containerID="9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363" Nov 24 13:38:23 crc kubenswrapper[4824]: E1124 13:38:23.527149 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363\": container with ID starting with 9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363 not found: ID does not exist" containerID="9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.527200 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363"} err="failed to get container status \"9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363\": rpc error: code = NotFound desc = could not find container \"9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363\": container with ID starting with 9d04dad2ec826e72bd0527d1cc4a62e6b2ca8e64c1d4b0ba513f6be449a57363 not found: ID does not exist" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.527230 4824 scope.go:117] "RemoveContainer" containerID="6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9" Nov 24 13:38:23 crc kubenswrapper[4824]: E1124 13:38:23.527612 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9\": container with ID starting with 6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9 not found: ID does not exist" containerID="6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9" Nov 24 13:38:23 crc kubenswrapper[4824]: I1124 13:38:23.527634 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9"} err="failed to get container status \"6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9\": rpc error: code = NotFound desc = could not find container \"6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9\": container with ID starting with 6381d22ca296a34c40c524936c1c5e885879e972c46d16f5e59039a198ac89e9 not found: ID does not exist" Nov 24 13:38:25 crc kubenswrapper[4824]: I1124 13:38:25.022184 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" path="/var/lib/kubelet/pods/a5c0f777-017e-4db2-8947-d27a71546d86/volumes" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.022717 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.109986 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6g4tc"] Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.110442 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" podUID="a2efd9de-42a7-49c7-a528-54919f22fbea" containerName="dnsmasq-dns" containerID="cri-o://25712dc18d9ce8eec007b3fbc507442babc42d3505dbc0dd5f8b31192ea2a0e0" gracePeriod=10 Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.301682 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d7677974f-hc64g"] Nov 24 13:38:27 crc kubenswrapper[4824]: E1124 13:38:27.305432 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="registry-server" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.305469 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="registry-server" Nov 24 13:38:27 crc kubenswrapper[4824]: E1124 13:38:27.305491 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="extract-content" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.305499 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="extract-content" Nov 24 13:38:27 crc kubenswrapper[4824]: E1124 13:38:27.305538 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="extract-utilities" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.305547 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="extract-utilities" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.305822 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c0f777-017e-4db2-8947-d27a71546d86" containerName="registry-server" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.307120 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.325254 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d7677974f-hc64g"] Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.394164 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-dns-svc\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.394284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8rkr\" (UniqueName: \"kubernetes.io/projected/0bec1542-f209-4da4-a540-9be4708ab637-kube-api-access-p8rkr\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.394398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-ovsdbserver-nb\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.394429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-config\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.394508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.394568 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-ovsdbserver-sb\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.394651 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-dns-swift-storage-0\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.444623 4824 generic.go:334] "Generic (PLEG): container finished" podID="a2efd9de-42a7-49c7-a528-54919f22fbea" containerID="25712dc18d9ce8eec007b3fbc507442babc42d3505dbc0dd5f8b31192ea2a0e0" exitCode=0 Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.444665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" event={"ID":"a2efd9de-42a7-49c7-a528-54919f22fbea","Type":"ContainerDied","Data":"25712dc18d9ce8eec007b3fbc507442babc42d3505dbc0dd5f8b31192ea2a0e0"} Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.496290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-ovsdbserver-sb\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.496395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-dns-swift-storage-0\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.497404 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-dns-svc\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.497791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-ovsdbserver-sb\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.497880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-dns-swift-storage-0\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.497922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-dns-svc\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.497952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8rkr\" (UniqueName: \"kubernetes.io/projected/0bec1542-f209-4da4-a540-9be4708ab637-kube-api-access-p8rkr\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.498005 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-ovsdbserver-nb\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.498026 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-config\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.498073 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.498572 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.499293 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-ovsdbserver-nb\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.499755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bec1542-f209-4da4-a540-9be4708ab637-config\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.535520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8rkr\" (UniqueName: \"kubernetes.io/projected/0bec1542-f209-4da4-a540-9be4708ab637-kube-api-access-p8rkr\") pod \"dnsmasq-dns-d7677974f-hc64g\" (UID: \"0bec1542-f209-4da4-a540-9be4708ab637\") " pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.668262 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:27 crc kubenswrapper[4824]: I1124 13:38:27.852525 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.013167 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-nb\") pod \"a2efd9de-42a7-49c7-a528-54919f22fbea\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.013227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf5qn\" (UniqueName: \"kubernetes.io/projected/a2efd9de-42a7-49c7-a528-54919f22fbea-kube-api-access-tf5qn\") pod \"a2efd9de-42a7-49c7-a528-54919f22fbea\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.013299 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-svc\") pod \"a2efd9de-42a7-49c7-a528-54919f22fbea\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.013407 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-config\") pod \"a2efd9de-42a7-49c7-a528-54919f22fbea\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.013434 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-sb\") pod \"a2efd9de-42a7-49c7-a528-54919f22fbea\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.013567 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-swift-storage-0\") pod \"a2efd9de-42a7-49c7-a528-54919f22fbea\" (UID: \"a2efd9de-42a7-49c7-a528-54919f22fbea\") " Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.026337 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2efd9de-42a7-49c7-a528-54919f22fbea-kube-api-access-tf5qn" (OuterVolumeSpecName: "kube-api-access-tf5qn") pod "a2efd9de-42a7-49c7-a528-54919f22fbea" (UID: "a2efd9de-42a7-49c7-a528-54919f22fbea"). InnerVolumeSpecName "kube-api-access-tf5qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.076593 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a2efd9de-42a7-49c7-a528-54919f22fbea" (UID: "a2efd9de-42a7-49c7-a528-54919f22fbea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.082513 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a2efd9de-42a7-49c7-a528-54919f22fbea" (UID: "a2efd9de-42a7-49c7-a528-54919f22fbea"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.084400 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a2efd9de-42a7-49c7-a528-54919f22fbea" (UID: "a2efd9de-42a7-49c7-a528-54919f22fbea"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.100345 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-config" (OuterVolumeSpecName: "config") pod "a2efd9de-42a7-49c7-a528-54919f22fbea" (UID: "a2efd9de-42a7-49c7-a528-54919f22fbea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.103169 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a2efd9de-42a7-49c7-a528-54919f22fbea" (UID: "a2efd9de-42a7-49c7-a528-54919f22fbea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.115693 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.115919 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.116083 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf5qn\" (UniqueName: \"kubernetes.io/projected/a2efd9de-42a7-49c7-a528-54919f22fbea-kube-api-access-tf5qn\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.116144 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.116194 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.116242 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2efd9de-42a7-49c7-a528-54919f22fbea-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.187147 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d7677974f-hc64g"] Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.460471 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" event={"ID":"a2efd9de-42a7-49c7-a528-54919f22fbea","Type":"ContainerDied","Data":"0659e61b7b260b13491ff10a57334371321c8b9f4636b720eff8591b6b4c1f56"} Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.460548 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-6g4tc" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.460565 4824 scope.go:117] "RemoveContainer" containerID="25712dc18d9ce8eec007b3fbc507442babc42d3505dbc0dd5f8b31192ea2a0e0" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.469525 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7677974f-hc64g" event={"ID":"0bec1542-f209-4da4-a540-9be4708ab637","Type":"ContainerStarted","Data":"bac0086545f03a14539c54dec3e5bdd4d171ae936085b7fe5aae03b6ac10301b"} Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.493555 4824 scope.go:117] "RemoveContainer" containerID="cb621c6b650ce5f8840fec7b5846413f285f346f675c4126d916f68833f321e6" Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.504521 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6g4tc"] Nov 24 13:38:28 crc kubenswrapper[4824]: I1124 13:38:28.513944 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6g4tc"] Nov 24 13:38:29 crc kubenswrapper[4824]: I1124 13:38:29.021649 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2efd9de-42a7-49c7-a528-54919f22fbea" path="/var/lib/kubelet/pods/a2efd9de-42a7-49c7-a528-54919f22fbea/volumes" Nov 24 13:38:29 crc kubenswrapper[4824]: I1124 13:38:29.481283 4824 generic.go:334] "Generic (PLEG): container finished" podID="0bec1542-f209-4da4-a540-9be4708ab637" containerID="55cb2845907589e1aa5bda757cda5d4b170f56c5b3471d65bd07043cd93c3f85" exitCode=0 Nov 24 13:38:29 crc kubenswrapper[4824]: I1124 13:38:29.481362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7677974f-hc64g" event={"ID":"0bec1542-f209-4da4-a540-9be4708ab637","Type":"ContainerDied","Data":"55cb2845907589e1aa5bda757cda5d4b170f56c5b3471d65bd07043cd93c3f85"} Nov 24 13:38:30 crc kubenswrapper[4824]: I1124 13:38:30.492729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7677974f-hc64g" event={"ID":"0bec1542-f209-4da4-a540-9be4708ab637","Type":"ContainerStarted","Data":"cbabf22023bab5d1379d80d363e68fea59fc9bc9aa845afa233aec1175b59ca2"} Nov 24 13:38:30 crc kubenswrapper[4824]: I1124 13:38:30.493278 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:30 crc kubenswrapper[4824]: I1124 13:38:30.515289 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d7677974f-hc64g" podStartSLOduration=3.51526627 podStartE2EDuration="3.51526627s" podCreationTimestamp="2025-11-24 13:38:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:38:30.509393391 +0000 UTC m=+1452.148932721" watchObservedRunningTime="2025-11-24 13:38:30.51526627 +0000 UTC m=+1452.154805600" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.363142 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jr4qt"] Nov 24 13:38:31 crc kubenswrapper[4824]: E1124 13:38:31.363626 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2efd9de-42a7-49c7-a528-54919f22fbea" containerName="dnsmasq-dns" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.363645 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2efd9de-42a7-49c7-a528-54919f22fbea" containerName="dnsmasq-dns" Nov 24 13:38:31 crc kubenswrapper[4824]: E1124 13:38:31.363677 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2efd9de-42a7-49c7-a528-54919f22fbea" containerName="init" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.363685 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2efd9de-42a7-49c7-a528-54919f22fbea" containerName="init" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.363986 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2efd9de-42a7-49c7-a528-54919f22fbea" containerName="dnsmasq-dns" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.365649 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.397563 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jr4qt"] Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.501798 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npnbk\" (UniqueName: \"kubernetes.io/projected/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-kube-api-access-npnbk\") pod \"certified-operators-jr4qt\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.502023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-catalog-content\") pod \"certified-operators-jr4qt\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.502067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-utilities\") pod \"certified-operators-jr4qt\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.603621 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-catalog-content\") pod \"certified-operators-jr4qt\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.603677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-utilities\") pod \"certified-operators-jr4qt\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.603728 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npnbk\" (UniqueName: \"kubernetes.io/projected/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-kube-api-access-npnbk\") pod \"certified-operators-jr4qt\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.604322 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-catalog-content\") pod \"certified-operators-jr4qt\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.604389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-utilities\") pod \"certified-operators-jr4qt\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.632965 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npnbk\" (UniqueName: \"kubernetes.io/projected/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-kube-api-access-npnbk\") pod \"certified-operators-jr4qt\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:31 crc kubenswrapper[4824]: I1124 13:38:31.706878 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:32 crc kubenswrapper[4824]: I1124 13:38:32.239712 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jr4qt"] Nov 24 13:38:32 crc kubenswrapper[4824]: I1124 13:38:32.516156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr4qt" event={"ID":"8c62fc1d-ce45-4ab6-8d6b-3569c648c110","Type":"ContainerStarted","Data":"fa5c83e34c7adfdb2a19ae352c2ae875e2be45cdb38bc4795f927767b1689fd8"} Nov 24 13:38:33 crc kubenswrapper[4824]: I1124 13:38:33.526737 4824 generic.go:334] "Generic (PLEG): container finished" podID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerID="0b569fad882da65e1f89b01252135d1af8a4c1c48dc1af5f9ec689f8f7dee040" exitCode=0 Nov 24 13:38:33 crc kubenswrapper[4824]: I1124 13:38:33.526786 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr4qt" event={"ID":"8c62fc1d-ce45-4ab6-8d6b-3569c648c110","Type":"ContainerDied","Data":"0b569fad882da65e1f89b01252135d1af8a4c1c48dc1af5f9ec689f8f7dee040"} Nov 24 13:38:36 crc kubenswrapper[4824]: I1124 13:38:36.557601 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr4qt" event={"ID":"8c62fc1d-ce45-4ab6-8d6b-3569c648c110","Type":"ContainerStarted","Data":"434c07b273fa8d0de8a9e5b3e6e54b13b06274c92268cd16f763a58c828cb842"} Nov 24 13:38:37 crc kubenswrapper[4824]: I1124 13:38:37.670908 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d7677974f-hc64g" Nov 24 13:38:37 crc kubenswrapper[4824]: I1124 13:38:37.765606 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dn5hh"] Nov 24 13:38:37 crc kubenswrapper[4824]: I1124 13:38:37.765875 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" podUID="675c8288-8692-4dd4-a66d-09e5447e349d" containerName="dnsmasq-dns" containerID="cri-o://432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc" gracePeriod=10 Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.361702 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.429516 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-openstack-edpm-ipam\") pod \"675c8288-8692-4dd4-a66d-09e5447e349d\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.429940 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-config\") pod \"675c8288-8692-4dd4-a66d-09e5447e349d\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.430097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-nb\") pod \"675c8288-8692-4dd4-a66d-09e5447e349d\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.430376 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-svc\") pod \"675c8288-8692-4dd4-a66d-09e5447e349d\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.430435 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-sb\") pod \"675c8288-8692-4dd4-a66d-09e5447e349d\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.430477 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2ss9\" (UniqueName: \"kubernetes.io/projected/675c8288-8692-4dd4-a66d-09e5447e349d-kube-api-access-c2ss9\") pod \"675c8288-8692-4dd4-a66d-09e5447e349d\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.430522 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-swift-storage-0\") pod \"675c8288-8692-4dd4-a66d-09e5447e349d\" (UID: \"675c8288-8692-4dd4-a66d-09e5447e349d\") " Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.467841 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/675c8288-8692-4dd4-a66d-09e5447e349d-kube-api-access-c2ss9" (OuterVolumeSpecName: "kube-api-access-c2ss9") pod "675c8288-8692-4dd4-a66d-09e5447e349d" (UID: "675c8288-8692-4dd4-a66d-09e5447e349d"). InnerVolumeSpecName "kube-api-access-c2ss9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.493376 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "675c8288-8692-4dd4-a66d-09e5447e349d" (UID: "675c8288-8692-4dd4-a66d-09e5447e349d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.538447 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2ss9\" (UniqueName: \"kubernetes.io/projected/675c8288-8692-4dd4-a66d-09e5447e349d-kube-api-access-c2ss9\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.538472 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.542409 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "675c8288-8692-4dd4-a66d-09e5447e349d" (UID: "675c8288-8692-4dd4-a66d-09e5447e349d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.548688 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-config" (OuterVolumeSpecName: "config") pod "675c8288-8692-4dd4-a66d-09e5447e349d" (UID: "675c8288-8692-4dd4-a66d-09e5447e349d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.564488 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "675c8288-8692-4dd4-a66d-09e5447e349d" (UID: "675c8288-8692-4dd4-a66d-09e5447e349d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.566264 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "675c8288-8692-4dd4-a66d-09e5447e349d" (UID: "675c8288-8692-4dd4-a66d-09e5447e349d"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.578942 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "675c8288-8692-4dd4-a66d-09e5447e349d" (UID: "675c8288-8692-4dd4-a66d-09e5447e349d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.581142 4824 generic.go:334] "Generic (PLEG): container finished" podID="675c8288-8692-4dd4-a66d-09e5447e349d" containerID="432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc" exitCode=0 Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.581185 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.581184 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" event={"ID":"675c8288-8692-4dd4-a66d-09e5447e349d","Type":"ContainerDied","Data":"432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc"} Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.581217 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dn5hh" event={"ID":"675c8288-8692-4dd4-a66d-09e5447e349d","Type":"ContainerDied","Data":"e9a5c6740a2276271c7fc65b3dc2b0221b8dedfde9cddfa505713b49db8c628e"} Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.581234 4824 scope.go:117] "RemoveContainer" containerID="432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.603203 4824 scope.go:117] "RemoveContainer" containerID="c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.619691 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dn5hh"] Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.628313 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dn5hh"] Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.628429 4824 scope.go:117] "RemoveContainer" containerID="432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc" Nov 24 13:38:38 crc kubenswrapper[4824]: E1124 13:38:38.628728 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc\": container with ID starting with 432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc not found: ID does not exist" containerID="432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.628757 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc"} err="failed to get container status \"432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc\": rpc error: code = NotFound desc = could not find container \"432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc\": container with ID starting with 432e52832d55b61e4dbbefc39796f7110723ce7c8d1ec6ff29b618e918f911dc not found: ID does not exist" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.628777 4824 scope.go:117] "RemoveContainer" containerID="c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f" Nov 24 13:38:38 crc kubenswrapper[4824]: E1124 13:38:38.629141 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f\": container with ID starting with c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f not found: ID does not exist" containerID="c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.629203 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f"} err="failed to get container status \"c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f\": rpc error: code = NotFound desc = could not find container \"c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f\": container with ID starting with c4b2a77768a3c05346b966da3d765825b0551257ecd1d8b389017ccfdf28be9f not found: ID does not exist" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.640646 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.640678 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.640690 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.640701 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:38 crc kubenswrapper[4824]: I1124 13:38:38.640714 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/675c8288-8692-4dd4-a66d-09e5447e349d-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:39 crc kubenswrapper[4824]: I1124 13:38:39.020294 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="675c8288-8692-4dd4-a66d-09e5447e349d" path="/var/lib/kubelet/pods/675c8288-8692-4dd4-a66d-09e5447e349d/volumes" Nov 24 13:38:40 crc kubenswrapper[4824]: I1124 13:38:40.605918 4824 generic.go:334] "Generic (PLEG): container finished" podID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerID="434c07b273fa8d0de8a9e5b3e6e54b13b06274c92268cd16f763a58c828cb842" exitCode=0 Nov 24 13:38:40 crc kubenswrapper[4824]: I1124 13:38:40.605986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr4qt" event={"ID":"8c62fc1d-ce45-4ab6-8d6b-3569c648c110","Type":"ContainerDied","Data":"434c07b273fa8d0de8a9e5b3e6e54b13b06274c92268cd16f763a58c828cb842"} Nov 24 13:38:40 crc kubenswrapper[4824]: I1124 13:38:40.788581 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:38:40 crc kubenswrapper[4824]: I1124 13:38:40.788679 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:38:42 crc kubenswrapper[4824]: I1124 13:38:42.626900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr4qt" event={"ID":"8c62fc1d-ce45-4ab6-8d6b-3569c648c110","Type":"ContainerStarted","Data":"bbec7cff655730f600e0aa2221054da899c4997d2ea1937dd73b610767fa5b2c"} Nov 24 13:38:42 crc kubenswrapper[4824]: I1124 13:38:42.649079 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jr4qt" podStartSLOduration=3.645960473 podStartE2EDuration="11.649056422s" podCreationTimestamp="2025-11-24 13:38:31 +0000 UTC" firstStartedPulling="2025-11-24 13:38:33.528909135 +0000 UTC m=+1455.168448465" lastFinishedPulling="2025-11-24 13:38:41.532005094 +0000 UTC m=+1463.171544414" observedRunningTime="2025-11-24 13:38:42.641844759 +0000 UTC m=+1464.281384069" watchObservedRunningTime="2025-11-24 13:38:42.649056422 +0000 UTC m=+1464.288595732" Nov 24 13:38:49 crc kubenswrapper[4824]: I1124 13:38:49.708189 4824 generic.go:334] "Generic (PLEG): container finished" podID="98ec6c14-59af-4b38-8e01-89ff157036ca" containerID="3b3364ac731f06487c715445e1125ca85845a9f9db1d5bf271b2b5a8615efc37" exitCode=0 Nov 24 13:38:49 crc kubenswrapper[4824]: I1124 13:38:49.708257 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"98ec6c14-59af-4b38-8e01-89ff157036ca","Type":"ContainerDied","Data":"3b3364ac731f06487c715445e1125ca85845a9f9db1d5bf271b2b5a8615efc37"} Nov 24 13:38:49 crc kubenswrapper[4824]: I1124 13:38:49.712302 4824 generic.go:334] "Generic (PLEG): container finished" podID="829719ca-f478-456b-9bbc-07cad8469731" containerID="7e24d3b66b28d5b61bfb4c84e469b05e234cc60e56ea2280447a2aeda9f13b2a" exitCode=0 Nov 24 13:38:49 crc kubenswrapper[4824]: I1124 13:38:49.712388 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"829719ca-f478-456b-9bbc-07cad8469731","Type":"ContainerDied","Data":"7e24d3b66b28d5b61bfb4c84e469b05e234cc60e56ea2280447a2aeda9f13b2a"} Nov 24 13:38:50 crc kubenswrapper[4824]: I1124 13:38:50.814444 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"829719ca-f478-456b-9bbc-07cad8469731","Type":"ContainerStarted","Data":"fa7fc797a38b399dbaadc1adb3ffb4624e033be97a73209c8780d165f07abe34"} Nov 24 13:38:51 crc kubenswrapper[4824]: I1124 13:38:51.707558 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:51 crc kubenswrapper[4824]: I1124 13:38:51.707922 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:51 crc kubenswrapper[4824]: I1124 13:38:51.762024 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:51 crc kubenswrapper[4824]: I1124 13:38:51.861737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"98ec6c14-59af-4b38-8e01-89ff157036ca","Type":"ContainerStarted","Data":"69340a64b7f65c77f4c8962ad36039e7cff7d283b12e872b0a3036466ef0258e"} Nov 24 13:38:51 crc kubenswrapper[4824]: I1124 13:38:51.862077 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 13:38:51 crc kubenswrapper[4824]: I1124 13:38:51.862618 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:38:51 crc kubenswrapper[4824]: I1124 13:38:51.914501 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.914483659 podStartE2EDuration="37.914483659s" podCreationTimestamp="2025-11-24 13:38:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:38:51.913433913 +0000 UTC m=+1473.552973223" watchObservedRunningTime="2025-11-24 13:38:51.914483659 +0000 UTC m=+1473.554022969" Nov 24 13:38:51 crc kubenswrapper[4824]: I1124 13:38:51.918657 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.918642375 podStartE2EDuration="37.918642375s" podCreationTimestamp="2025-11-24 13:38:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:38:51.88932022 +0000 UTC m=+1473.528859530" watchObservedRunningTime="2025-11-24 13:38:51.918642375 +0000 UTC m=+1473.558181695" Nov 24 13:38:51 crc kubenswrapper[4824]: I1124 13:38:51.923412 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:52 crc kubenswrapper[4824]: I1124 13:38:52.002575 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jr4qt"] Nov 24 13:38:53 crc kubenswrapper[4824]: I1124 13:38:53.876905 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jr4qt" podUID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerName="registry-server" containerID="cri-o://bbec7cff655730f600e0aa2221054da899c4997d2ea1937dd73b610767fa5b2c" gracePeriod=2 Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.886542 4824 generic.go:334] "Generic (PLEG): container finished" podID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerID="bbec7cff655730f600e0aa2221054da899c4997d2ea1937dd73b610767fa5b2c" exitCode=0 Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.886592 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr4qt" event={"ID":"8c62fc1d-ce45-4ab6-8d6b-3569c648c110","Type":"ContainerDied","Data":"bbec7cff655730f600e0aa2221054da899c4997d2ea1937dd73b610767fa5b2c"} Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.886850 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr4qt" event={"ID":"8c62fc1d-ce45-4ab6-8d6b-3569c648c110","Type":"ContainerDied","Data":"fa5c83e34c7adfdb2a19ae352c2ae875e2be45cdb38bc4795f927767b1689fd8"} Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.886866 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa5c83e34c7adfdb2a19ae352c2ae875e2be45cdb38bc4795f927767b1689fd8" Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.922174 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.963351 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npnbk\" (UniqueName: \"kubernetes.io/projected/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-kube-api-access-npnbk\") pod \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.963453 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-utilities\") pod \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.963525 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-catalog-content\") pod \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\" (UID: \"8c62fc1d-ce45-4ab6-8d6b-3569c648c110\") " Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.976328 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-utilities" (OuterVolumeSpecName: "utilities") pod "8c62fc1d-ce45-4ab6-8d6b-3569c648c110" (UID: "8c62fc1d-ce45-4ab6-8d6b-3569c648c110"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:38:54 crc kubenswrapper[4824]: I1124 13:38:54.990977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-kube-api-access-npnbk" (OuterVolumeSpecName: "kube-api-access-npnbk") pod "8c62fc1d-ce45-4ab6-8d6b-3569c648c110" (UID: "8c62fc1d-ce45-4ab6-8d6b-3569c648c110"). InnerVolumeSpecName "kube-api-access-npnbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:38:55 crc kubenswrapper[4824]: I1124 13:38:55.053335 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c62fc1d-ce45-4ab6-8d6b-3569c648c110" (UID: "8c62fc1d-ce45-4ab6-8d6b-3569c648c110"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:38:55 crc kubenswrapper[4824]: I1124 13:38:55.066102 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npnbk\" (UniqueName: \"kubernetes.io/projected/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-kube-api-access-npnbk\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:55 crc kubenswrapper[4824]: I1124 13:38:55.066134 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:55 crc kubenswrapper[4824]: I1124 13:38:55.066145 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c62fc1d-ce45-4ab6-8d6b-3569c648c110-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:38:55 crc kubenswrapper[4824]: I1124 13:38:55.895014 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr4qt" Nov 24 13:38:55 crc kubenswrapper[4824]: I1124 13:38:55.932907 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jr4qt"] Nov 24 13:38:55 crc kubenswrapper[4824]: I1124 13:38:55.953414 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jr4qt"] Nov 24 13:38:57 crc kubenswrapper[4824]: I1124 13:38:57.026987 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" path="/var/lib/kubelet/pods/8c62fc1d-ce45-4ab6-8d6b-3569c648c110/volumes" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.560687 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh"] Nov 24 13:39:01 crc kubenswrapper[4824]: E1124 13:39:01.561635 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="675c8288-8692-4dd4-a66d-09e5447e349d" containerName="dnsmasq-dns" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.561650 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="675c8288-8692-4dd4-a66d-09e5447e349d" containerName="dnsmasq-dns" Nov 24 13:39:01 crc kubenswrapper[4824]: E1124 13:39:01.561678 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerName="extract-utilities" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.561687 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerName="extract-utilities" Nov 24 13:39:01 crc kubenswrapper[4824]: E1124 13:39:01.561720 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="675c8288-8692-4dd4-a66d-09e5447e349d" containerName="init" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.561728 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="675c8288-8692-4dd4-a66d-09e5447e349d" containerName="init" Nov 24 13:39:01 crc kubenswrapper[4824]: E1124 13:39:01.561741 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerName="registry-server" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.561751 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerName="registry-server" Nov 24 13:39:01 crc kubenswrapper[4824]: E1124 13:39:01.561771 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerName="extract-content" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.561779 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerName="extract-content" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.562034 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="675c8288-8692-4dd4-a66d-09e5447e349d" containerName="dnsmasq-dns" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.562058 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c62fc1d-ce45-4ab6-8d6b-3569c648c110" containerName="registry-server" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.562796 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.565514 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.567359 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.567373 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.567870 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.590111 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh"] Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.597480 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6ks4\" (UniqueName: \"kubernetes.io/projected/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-kube-api-access-g6ks4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.597550 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.597639 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.597713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.699636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6ks4\" (UniqueName: \"kubernetes.io/projected/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-kube-api-access-g6ks4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.699954 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.700003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.700067 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.705680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.706524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.710439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.745449 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6ks4\" (UniqueName: \"kubernetes.io/projected/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-kube-api-access-g6ks4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:01 crc kubenswrapper[4824]: I1124 13:39:01.886217 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:03 crc kubenswrapper[4824]: I1124 13:39:03.559749 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh"] Nov 24 13:39:03 crc kubenswrapper[4824]: W1124 13:39:03.565665 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a2dbe7f_7854_4dc7_9511_bba8a8c81a08.slice/crio-19032564c95404a08c84a92009303304176379dca37102a1323dceabf67129d5 WatchSource:0}: Error finding container 19032564c95404a08c84a92009303304176379dca37102a1323dceabf67129d5: Status 404 returned error can't find the container with id 19032564c95404a08c84a92009303304176379dca37102a1323dceabf67129d5 Nov 24 13:39:03 crc kubenswrapper[4824]: I1124 13:39:03.967235 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" event={"ID":"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08","Type":"ContainerStarted","Data":"19032564c95404a08c84a92009303304176379dca37102a1323dceabf67129d5"} Nov 24 13:39:04 crc kubenswrapper[4824]: I1124 13:39:04.844658 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="98ec6c14-59af-4b38-8e01-89ff157036ca" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.206:5671: connect: connection refused" Nov 24 13:39:04 crc kubenswrapper[4824]: I1124 13:39:04.846432 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="829719ca-f478-456b-9bbc-07cad8469731" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.207:5671: connect: connection refused" Nov 24 13:39:10 crc kubenswrapper[4824]: I1124 13:39:10.789218 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:39:10 crc kubenswrapper[4824]: I1124 13:39:10.789874 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:39:10 crc kubenswrapper[4824]: I1124 13:39:10.789955 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:39:10 crc kubenswrapper[4824]: I1124 13:39:10.791015 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1981738b6eecfa8db6367aadef18ef52e9a453a009ef7bb2c198c0fe6a176d1"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:39:10 crc kubenswrapper[4824]: I1124 13:39:10.791197 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://b1981738b6eecfa8db6367aadef18ef52e9a453a009ef7bb2c198c0fe6a176d1" gracePeriod=600 Nov 24 13:39:11 crc kubenswrapper[4824]: I1124 13:39:11.066866 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="b1981738b6eecfa8db6367aadef18ef52e9a453a009ef7bb2c198c0fe6a176d1" exitCode=0 Nov 24 13:39:11 crc kubenswrapper[4824]: I1124 13:39:11.066921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"b1981738b6eecfa8db6367aadef18ef52e9a453a009ef7bb2c198c0fe6a176d1"} Nov 24 13:39:11 crc kubenswrapper[4824]: I1124 13:39:11.066952 4824 scope.go:117] "RemoveContainer" containerID="7023e2fedbc8a159e146d80c57d0a65acc7f7defe0350732f194124c1f095b6e" Nov 24 13:39:14 crc kubenswrapper[4824]: I1124 13:39:14.845015 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 13:39:14 crc kubenswrapper[4824]: I1124 13:39:14.846011 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 13:39:20 crc kubenswrapper[4824]: E1124 13:39:20.144880 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:3b0bb6b72928679854fd8df2180c23d5c3286974" Nov 24 13:39:20 crc kubenswrapper[4824]: E1124 13:39:20.145473 4824 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:3b0bb6b72928679854fd8df2180c23d5c3286974" Nov 24 13:39:20 crc kubenswrapper[4824]: E1124 13:39:20.152824 4824 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 24 13:39:20 crc kubenswrapper[4824]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:3b0bb6b72928679854fd8df2180c23d5c3286974,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Nov 24 13:39:20 crc kubenswrapper[4824]: - hosts: all Nov 24 13:39:20 crc kubenswrapper[4824]: strategy: linear Nov 24 13:39:20 crc kubenswrapper[4824]: tasks: Nov 24 13:39:20 crc kubenswrapper[4824]: - name: Enable podified-repos Nov 24 13:39:20 crc kubenswrapper[4824]: become: true Nov 24 13:39:20 crc kubenswrapper[4824]: ansible.builtin.shell: | Nov 24 13:39:20 crc kubenswrapper[4824]: set -euxo pipefail Nov 24 13:39:20 crc kubenswrapper[4824]: pushd /var/tmp Nov 24 13:39:20 crc kubenswrapper[4824]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Nov 24 13:39:20 crc kubenswrapper[4824]: pushd repo-setup-main Nov 24 13:39:20 crc kubenswrapper[4824]: python3 -m venv ./venv Nov 24 13:39:20 crc kubenswrapper[4824]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Nov 24 13:39:20 crc kubenswrapper[4824]: ./venv/bin/repo-setup current-podified -b antelope Nov 24 13:39:20 crc kubenswrapper[4824]: popd Nov 24 13:39:20 crc kubenswrapper[4824]: rm -rf repo-setup-main Nov 24 13:39:20 crc kubenswrapper[4824]: Nov 24 13:39:20 crc kubenswrapper[4824]: Nov 24 13:39:20 crc kubenswrapper[4824]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Nov 24 13:39:20 crc kubenswrapper[4824]: edpm_override_hosts: openstack-edpm-ipam Nov 24 13:39:20 crc kubenswrapper[4824]: edpm_service_type: repo-setup Nov 24 13:39:20 crc kubenswrapper[4824]: Nov 24 13:39:20 crc kubenswrapper[4824]: Nov 24 13:39:20 crc kubenswrapper[4824]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g6ks4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh_openstack(1a2dbe7f-7854-4dc7-9511-bba8a8c81a08): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 24 13:39:20 crc kubenswrapper[4824]: > logger="UnhandledError" Nov 24 13:39:20 crc kubenswrapper[4824]: E1124 13:39:20.154671 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" podUID="1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" Nov 24 13:39:20 crc kubenswrapper[4824]: E1124 13:39:20.196057 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:3b0bb6b72928679854fd8df2180c23d5c3286974\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" podUID="1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" Nov 24 13:39:21 crc kubenswrapper[4824]: I1124 13:39:21.209719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80"} Nov 24 13:39:22 crc kubenswrapper[4824]: I1124 13:39:22.017720 4824 scope.go:117] "RemoveContainer" containerID="260fbee5605e0180a413e46b64f14f0374ba776b4d5d74aada3ce1635c31a16b" Nov 24 13:39:35 crc kubenswrapper[4824]: I1124 13:39:35.210746 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:39:36 crc kubenswrapper[4824]: I1124 13:39:36.385582 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" event={"ID":"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08","Type":"ContainerStarted","Data":"8fa471684eabda1519b0973ef78b6b8dff10071da11e77fc06a49ddeb73c2b23"} Nov 24 13:39:36 crc kubenswrapper[4824]: I1124 13:39:36.417484 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" podStartSLOduration=3.7771778019999998 podStartE2EDuration="35.417461676s" podCreationTimestamp="2025-11-24 13:39:01 +0000 UTC" firstStartedPulling="2025-11-24 13:39:03.568001085 +0000 UTC m=+1485.207540405" lastFinishedPulling="2025-11-24 13:39:35.208284969 +0000 UTC m=+1516.847824279" observedRunningTime="2025-11-24 13:39:36.406230741 +0000 UTC m=+1518.045770051" watchObservedRunningTime="2025-11-24 13:39:36.417461676 +0000 UTC m=+1518.057000996" Nov 24 13:39:49 crc kubenswrapper[4824]: I1124 13:39:49.534713 4824 generic.go:334] "Generic (PLEG): container finished" podID="1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" containerID="8fa471684eabda1519b0973ef78b6b8dff10071da11e77fc06a49ddeb73c2b23" exitCode=0 Nov 24 13:39:49 crc kubenswrapper[4824]: I1124 13:39:49.534849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" event={"ID":"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08","Type":"ContainerDied","Data":"8fa471684eabda1519b0973ef78b6b8dff10071da11e77fc06a49ddeb73c2b23"} Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.006180 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.112944 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6ks4\" (UniqueName: \"kubernetes.io/projected/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-kube-api-access-g6ks4\") pod \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.115161 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-inventory\") pod \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.115215 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-repo-setup-combined-ca-bundle\") pod \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.115290 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-ssh-key\") pod \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\" (UID: \"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08\") " Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.123195 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-kube-api-access-g6ks4" (OuterVolumeSpecName: "kube-api-access-g6ks4") pod "1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" (UID: "1a2dbe7f-7854-4dc7-9511-bba8a8c81a08"). InnerVolumeSpecName "kube-api-access-g6ks4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.123189 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" (UID: "1a2dbe7f-7854-4dc7-9511-bba8a8c81a08"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.145820 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" (UID: "1a2dbe7f-7854-4dc7-9511-bba8a8c81a08"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.148306 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-inventory" (OuterVolumeSpecName: "inventory") pod "1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" (UID: "1a2dbe7f-7854-4dc7-9511-bba8a8c81a08"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.219518 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.219549 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.219558 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.219566 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6ks4\" (UniqueName: \"kubernetes.io/projected/1a2dbe7f-7854-4dc7-9511-bba8a8c81a08-kube-api-access-g6ks4\") on node \"crc\" DevicePath \"\"" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.554873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" event={"ID":"1a2dbe7f-7854-4dc7-9511-bba8a8c81a08","Type":"ContainerDied","Data":"19032564c95404a08c84a92009303304176379dca37102a1323dceabf67129d5"} Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.554905 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.554911 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19032564c95404a08c84a92009303304176379dca37102a1323dceabf67129d5" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.651228 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr"] Nov 24 13:39:51 crc kubenswrapper[4824]: E1124 13:39:51.651675 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.651698 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.651954 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2dbe7f-7854-4dc7-9511-bba8a8c81a08" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.652628 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.657600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.657919 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.658156 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.658341 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.671747 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr"] Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.830071 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-94hhr\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.830410 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb6fz\" (UniqueName: \"kubernetes.io/projected/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-kube-api-access-vb6fz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-94hhr\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.831020 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-94hhr\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.932731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-94hhr\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.933113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-94hhr\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.933191 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb6fz\" (UniqueName: \"kubernetes.io/projected/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-kube-api-access-vb6fz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-94hhr\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.939600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-94hhr\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.940260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-94hhr\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.957764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb6fz\" (UniqueName: \"kubernetes.io/projected/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-kube-api-access-vb6fz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-94hhr\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:51 crc kubenswrapper[4824]: I1124 13:39:51.993400 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:52 crc kubenswrapper[4824]: I1124 13:39:52.522108 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr"] Nov 24 13:39:52 crc kubenswrapper[4824]: W1124 13:39:52.531503 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53c68f20_6a1f_449e_a54e_aa1df1bf7d32.slice/crio-2bdc106b06c9a6e35cbbd74bb7140277a818c0e8952da3380ff80b51abc5842a WatchSource:0}: Error finding container 2bdc106b06c9a6e35cbbd74bb7140277a818c0e8952da3380ff80b51abc5842a: Status 404 returned error can't find the container with id 2bdc106b06c9a6e35cbbd74bb7140277a818c0e8952da3380ff80b51abc5842a Nov 24 13:39:52 crc kubenswrapper[4824]: I1124 13:39:52.564142 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" event={"ID":"53c68f20-6a1f-449e-a54e-aa1df1bf7d32","Type":"ContainerStarted","Data":"2bdc106b06c9a6e35cbbd74bb7140277a818c0e8952da3380ff80b51abc5842a"} Nov 24 13:39:53 crc kubenswrapper[4824]: I1124 13:39:53.574725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" event={"ID":"53c68f20-6a1f-449e-a54e-aa1df1bf7d32","Type":"ContainerStarted","Data":"e5d370407f27af4813fe7128e92d1a45892900a5f3942bd9890cb1f6617001e1"} Nov 24 13:39:53 crc kubenswrapper[4824]: I1124 13:39:53.601817 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" podStartSLOduration=2.395876114 podStartE2EDuration="2.601785714s" podCreationTimestamp="2025-11-24 13:39:51 +0000 UTC" firstStartedPulling="2025-11-24 13:39:52.535424942 +0000 UTC m=+1534.174964252" lastFinishedPulling="2025-11-24 13:39:52.741334542 +0000 UTC m=+1534.380873852" observedRunningTime="2025-11-24 13:39:53.59099949 +0000 UTC m=+1535.230538800" watchObservedRunningTime="2025-11-24 13:39:53.601785714 +0000 UTC m=+1535.241325024" Nov 24 13:39:55 crc kubenswrapper[4824]: I1124 13:39:55.597089 4824 generic.go:334] "Generic (PLEG): container finished" podID="53c68f20-6a1f-449e-a54e-aa1df1bf7d32" containerID="e5d370407f27af4813fe7128e92d1a45892900a5f3942bd9890cb1f6617001e1" exitCode=0 Nov 24 13:39:55 crc kubenswrapper[4824]: I1124 13:39:55.597173 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" event={"ID":"53c68f20-6a1f-449e-a54e-aa1df1bf7d32","Type":"ContainerDied","Data":"e5d370407f27af4813fe7128e92d1a45892900a5f3942bd9890cb1f6617001e1"} Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.030217 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.144636 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-ssh-key\") pod \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.144675 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb6fz\" (UniqueName: \"kubernetes.io/projected/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-kube-api-access-vb6fz\") pod \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.144901 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-inventory\") pod \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\" (UID: \"53c68f20-6a1f-449e-a54e-aa1df1bf7d32\") " Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.155331 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-kube-api-access-vb6fz" (OuterVolumeSpecName: "kube-api-access-vb6fz") pod "53c68f20-6a1f-449e-a54e-aa1df1bf7d32" (UID: "53c68f20-6a1f-449e-a54e-aa1df1bf7d32"). InnerVolumeSpecName "kube-api-access-vb6fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.205169 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-inventory" (OuterVolumeSpecName: "inventory") pod "53c68f20-6a1f-449e-a54e-aa1df1bf7d32" (UID: "53c68f20-6a1f-449e-a54e-aa1df1bf7d32"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.205880 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "53c68f20-6a1f-449e-a54e-aa1df1bf7d32" (UID: "53c68f20-6a1f-449e-a54e-aa1df1bf7d32"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.247496 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.247536 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.247547 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb6fz\" (UniqueName: \"kubernetes.io/projected/53c68f20-6a1f-449e-a54e-aa1df1bf7d32-kube-api-access-vb6fz\") on node \"crc\" DevicePath \"\"" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.612696 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" event={"ID":"53c68f20-6a1f-449e-a54e-aa1df1bf7d32","Type":"ContainerDied","Data":"2bdc106b06c9a6e35cbbd74bb7140277a818c0e8952da3380ff80b51abc5842a"} Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.612736 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bdc106b06c9a6e35cbbd74bb7140277a818c0e8952da3380ff80b51abc5842a" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.612819 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-94hhr" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.703012 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs"] Nov 24 13:39:57 crc kubenswrapper[4824]: E1124 13:39:57.703512 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c68f20-6a1f-449e-a54e-aa1df1bf7d32" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.703535 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c68f20-6a1f-449e-a54e-aa1df1bf7d32" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.703789 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c68f20-6a1f-449e-a54e-aa1df1bf7d32" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.704793 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.708438 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.710028 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.710112 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.711015 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.712073 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs"] Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.759373 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.759419 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2vf5\" (UniqueName: \"kubernetes.io/projected/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-kube-api-access-m2vf5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.759495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.759525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.861154 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.861203 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2vf5\" (UniqueName: \"kubernetes.io/projected/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-kube-api-access-m2vf5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.861309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.861381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.866916 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.866929 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.868237 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:57 crc kubenswrapper[4824]: I1124 13:39:57.880519 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2vf5\" (UniqueName: \"kubernetes.io/projected/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-kube-api-access-m2vf5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:58 crc kubenswrapper[4824]: I1124 13:39:58.034115 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:39:58 crc kubenswrapper[4824]: I1124 13:39:58.561905 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs"] Nov 24 13:39:58 crc kubenswrapper[4824]: I1124 13:39:58.622525 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" event={"ID":"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7","Type":"ContainerStarted","Data":"7e5e8bbf0e27e6560abe1211e6efea39f88dd6a3cb96c49120d5eb77bfbcdbb9"} Nov 24 13:39:59 crc kubenswrapper[4824]: I1124 13:39:59.633058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" event={"ID":"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7","Type":"ContainerStarted","Data":"dba1dac792c06b61f5bee5f4fcb20f3fcea0c9403c2ba4da3eb01693676fd5c5"} Nov 24 13:39:59 crc kubenswrapper[4824]: I1124 13:39:59.657516 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" podStartSLOduration=2.471452271 podStartE2EDuration="2.657493416s" podCreationTimestamp="2025-11-24 13:39:57 +0000 UTC" firstStartedPulling="2025-11-24 13:39:58.562451606 +0000 UTC m=+1540.201990926" lastFinishedPulling="2025-11-24 13:39:58.748492761 +0000 UTC m=+1540.388032071" observedRunningTime="2025-11-24 13:39:59.652785196 +0000 UTC m=+1541.292324506" watchObservedRunningTime="2025-11-24 13:39:59.657493416 +0000 UTC m=+1541.297032746" Nov 24 13:40:22 crc kubenswrapper[4824]: I1124 13:40:22.168088 4824 scope.go:117] "RemoveContainer" containerID="c8a1a3455e6f6c0c3b5919f99427cbab96f94a517d88a097709de0ae63e2e964" Nov 24 13:41:40 crc kubenswrapper[4824]: I1124 13:41:40.799364 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:41:40 crc kubenswrapper[4824]: I1124 13:41:40.799854 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:42:10 crc kubenswrapper[4824]: I1124 13:42:10.787995 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:42:10 crc kubenswrapper[4824]: I1124 13:42:10.788533 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.042889 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7xskl"] Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.045566 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.060323 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7xskl"] Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.084196 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-catalog-content\") pod \"community-operators-7xskl\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.084260 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbjvv\" (UniqueName: \"kubernetes.io/projected/5488c504-7399-4f08-bcb7-d01a64de5da5-kube-api-access-vbjvv\") pod \"community-operators-7xskl\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.084300 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-utilities\") pod \"community-operators-7xskl\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.185533 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-catalog-content\") pod \"community-operators-7xskl\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.185584 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbjvv\" (UniqueName: \"kubernetes.io/projected/5488c504-7399-4f08-bcb7-d01a64de5da5-kube-api-access-vbjvv\") pod \"community-operators-7xskl\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.185617 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-utilities\") pod \"community-operators-7xskl\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.186113 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-utilities\") pod \"community-operators-7xskl\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.186248 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-catalog-content\") pod \"community-operators-7xskl\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.207060 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbjvv\" (UniqueName: \"kubernetes.io/projected/5488c504-7399-4f08-bcb7-d01a64de5da5-kube-api-access-vbjvv\") pod \"community-operators-7xskl\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.379638 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:30 crc kubenswrapper[4824]: I1124 13:42:30.818543 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7xskl"] Nov 24 13:42:31 crc kubenswrapper[4824]: I1124 13:42:31.229113 4824 generic.go:334] "Generic (PLEG): container finished" podID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerID="a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283" exitCode=0 Nov 24 13:42:31 crc kubenswrapper[4824]: I1124 13:42:31.229253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xskl" event={"ID":"5488c504-7399-4f08-bcb7-d01a64de5da5","Type":"ContainerDied","Data":"a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283"} Nov 24 13:42:31 crc kubenswrapper[4824]: I1124 13:42:31.229359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xskl" event={"ID":"5488c504-7399-4f08-bcb7-d01a64de5da5","Type":"ContainerStarted","Data":"d9dffdccac55798d874da23933801070aac5ef8918c6c668545170345cba171e"} Nov 24 13:42:31 crc kubenswrapper[4824]: I1124 13:42:31.231983 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:42:40 crc kubenswrapper[4824]: I1124 13:42:40.787551 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:42:40 crc kubenswrapper[4824]: I1124 13:42:40.788106 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:42:40 crc kubenswrapper[4824]: I1124 13:42:40.788155 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:42:40 crc kubenswrapper[4824]: I1124 13:42:40.790464 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:42:40 crc kubenswrapper[4824]: I1124 13:42:40.790553 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" gracePeriod=600 Nov 24 13:42:41 crc kubenswrapper[4824]: I1124 13:42:41.373747 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" exitCode=0 Nov 24 13:42:41 crc kubenswrapper[4824]: I1124 13:42:41.373841 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80"} Nov 24 13:42:41 crc kubenswrapper[4824]: I1124 13:42:41.374162 4824 scope.go:117] "RemoveContainer" containerID="b1981738b6eecfa8db6367aadef18ef52e9a453a009ef7bb2c198c0fe6a176d1" Nov 24 13:42:42 crc kubenswrapper[4824]: E1124 13:42:42.342142 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:42:42 crc kubenswrapper[4824]: I1124 13:42:42.385085 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:42:42 crc kubenswrapper[4824]: E1124 13:42:42.385376 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:42:44 crc kubenswrapper[4824]: I1124 13:42:44.409619 4824 generic.go:334] "Generic (PLEG): container finished" podID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerID="f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd" exitCode=0 Nov 24 13:42:44 crc kubenswrapper[4824]: I1124 13:42:44.409725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xskl" event={"ID":"5488c504-7399-4f08-bcb7-d01a64de5da5","Type":"ContainerDied","Data":"f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd"} Nov 24 13:42:46 crc kubenswrapper[4824]: I1124 13:42:46.432539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xskl" event={"ID":"5488c504-7399-4f08-bcb7-d01a64de5da5","Type":"ContainerStarted","Data":"5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10"} Nov 24 13:42:46 crc kubenswrapper[4824]: I1124 13:42:46.464122 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7xskl" podStartSLOduration=2.394944206 podStartE2EDuration="16.464096821s" podCreationTimestamp="2025-11-24 13:42:30 +0000 UTC" firstStartedPulling="2025-11-24 13:42:31.231771389 +0000 UTC m=+1692.871310699" lastFinishedPulling="2025-11-24 13:42:45.300924004 +0000 UTC m=+1706.940463314" observedRunningTime="2025-11-24 13:42:46.45143317 +0000 UTC m=+1708.090972480" watchObservedRunningTime="2025-11-24 13:42:46.464096821 +0000 UTC m=+1708.103636141" Nov 24 13:42:50 crc kubenswrapper[4824]: I1124 13:42:50.380083 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:50 crc kubenswrapper[4824]: I1124 13:42:50.380737 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:50 crc kubenswrapper[4824]: I1124 13:42:50.451180 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:50 crc kubenswrapper[4824]: I1124 13:42:50.518772 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7xskl" Nov 24 13:42:50 crc kubenswrapper[4824]: I1124 13:42:50.694626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7xskl"] Nov 24 13:42:50 crc kubenswrapper[4824]: I1124 13:42:50.767398 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gbfbm"] Nov 24 13:42:50 crc kubenswrapper[4824]: I1124 13:42:50.767623 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gbfbm" podUID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerName="registry-server" containerID="cri-o://b0e094566c4fad9761e0c1c1f4af752c12e396e652e890e17bccf3c2c6202281" gracePeriod=2 Nov 24 13:42:53 crc kubenswrapper[4824]: I1124 13:42:53.528839 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gbfbm_dec682c6-9c93-4e5f-a595-51a31d308c46/registry-server/0.log" Nov 24 13:42:53 crc kubenswrapper[4824]: I1124 13:42:53.530573 4824 generic.go:334] "Generic (PLEG): container finished" podID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerID="b0e094566c4fad9761e0c1c1f4af752c12e396e652e890e17bccf3c2c6202281" exitCode=137 Nov 24 13:42:53 crc kubenswrapper[4824]: I1124 13:42:53.530608 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbfbm" event={"ID":"dec682c6-9c93-4e5f-a595-51a31d308c46","Type":"ContainerDied","Data":"b0e094566c4fad9761e0c1c1f4af752c12e396e652e890e17bccf3c2c6202281"} Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.017730 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:42:55 crc kubenswrapper[4824]: E1124 13:42:55.018499 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.106849 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2gs6v"] Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.119519 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-b3c8-account-create-n8x6v"] Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.132081 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-lsnff"] Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.145569 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2gs6v"] Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.162725 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-b3c8-account-create-n8x6v"] Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.173467 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-lsnff"] Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.229305 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gbfbm_dec682c6-9c93-4e5f-a595-51a31d308c46/registry-server/0.log" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.229947 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.373611 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwlpx\" (UniqueName: \"kubernetes.io/projected/dec682c6-9c93-4e5f-a595-51a31d308c46-kube-api-access-vwlpx\") pod \"dec682c6-9c93-4e5f-a595-51a31d308c46\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.373824 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-catalog-content\") pod \"dec682c6-9c93-4e5f-a595-51a31d308c46\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.374030 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-utilities\") pod \"dec682c6-9c93-4e5f-a595-51a31d308c46\" (UID: \"dec682c6-9c93-4e5f-a595-51a31d308c46\") " Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.374616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-utilities" (OuterVolumeSpecName: "utilities") pod "dec682c6-9c93-4e5f-a595-51a31d308c46" (UID: "dec682c6-9c93-4e5f-a595-51a31d308c46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.386134 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dec682c6-9c93-4e5f-a595-51a31d308c46-kube-api-access-vwlpx" (OuterVolumeSpecName: "kube-api-access-vwlpx") pod "dec682c6-9c93-4e5f-a595-51a31d308c46" (UID: "dec682c6-9c93-4e5f-a595-51a31d308c46"). InnerVolumeSpecName "kube-api-access-vwlpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.421096 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dec682c6-9c93-4e5f-a595-51a31d308c46" (UID: "dec682c6-9c93-4e5f-a595-51a31d308c46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.477014 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.477327 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dec682c6-9c93-4e5f-a595-51a31d308c46-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.477344 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwlpx\" (UniqueName: \"kubernetes.io/projected/dec682c6-9c93-4e5f-a595-51a31d308c46-kube-api-access-vwlpx\") on node \"crc\" DevicePath \"\"" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.650235 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gbfbm_dec682c6-9c93-4e5f-a595-51a31d308c46/registry-server/0.log" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.651029 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbfbm" event={"ID":"dec682c6-9c93-4e5f-a595-51a31d308c46","Type":"ContainerDied","Data":"aa518deca1a6bda27431e38794ecd38b8bc36bfd6221e7033b5f67d6afb0d946"} Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.651083 4824 scope.go:117] "RemoveContainer" containerID="b0e094566c4fad9761e0c1c1f4af752c12e396e652e890e17bccf3c2c6202281" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.651237 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbfbm" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.699071 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gbfbm"] Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.707543 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gbfbm"] Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.714535 4824 scope.go:117] "RemoveContainer" containerID="09dae5879834cb0d54d3c312d83055e60cb90aa3a5bbade3088d82fdc321578e" Nov 24 13:42:55 crc kubenswrapper[4824]: I1124 13:42:55.743569 4824 scope.go:117] "RemoveContainer" containerID="7214804e700748f3eb18b45c62e362e34da34042d6361bcd4430216536f37997" Nov 24 13:42:56 crc kubenswrapper[4824]: I1124 13:42:56.039868 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-3410-account-create-c74k2"] Nov 24 13:42:56 crc kubenswrapper[4824]: I1124 13:42:56.049535 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-3410-account-create-c74k2"] Nov 24 13:42:57 crc kubenswrapper[4824]: I1124 13:42:57.020535 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ac49a1-753f-46bf-bddd-410c28fe864e" path="/var/lib/kubelet/pods/01ac49a1-753f-46bf-bddd-410c28fe864e/volumes" Nov 24 13:42:57 crc kubenswrapper[4824]: I1124 13:42:57.021371 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="827adb53-2cb6-46ec-be78-6d25ced71cff" path="/var/lib/kubelet/pods/827adb53-2cb6-46ec-be78-6d25ced71cff/volumes" Nov 24 13:42:57 crc kubenswrapper[4824]: I1124 13:42:57.021955 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbba9a43-445e-4657-9a7b-35cd2d2dc405" path="/var/lib/kubelet/pods/cbba9a43-445e-4657-9a7b-35cd2d2dc405/volumes" Nov 24 13:42:57 crc kubenswrapper[4824]: I1124 13:42:57.022544 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dec682c6-9c93-4e5f-a595-51a31d308c46" path="/var/lib/kubelet/pods/dec682c6-9c93-4e5f-a595-51a31d308c46/volumes" Nov 24 13:42:57 crc kubenswrapper[4824]: I1124 13:42:57.023684 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e25c9bc5-d0fc-48e4-9579-59b360e49731" path="/var/lib/kubelet/pods/e25c9bc5-d0fc-48e4-9579-59b360e49731/volumes" Nov 24 13:43:03 crc kubenswrapper[4824]: I1124 13:43:03.057322 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-scfqc"] Nov 24 13:43:03 crc kubenswrapper[4824]: I1124 13:43:03.074284 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6389-account-create-2pp9j"] Nov 24 13:43:03 crc kubenswrapper[4824]: I1124 13:43:03.091945 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-scfqc"] Nov 24 13:43:03 crc kubenswrapper[4824]: I1124 13:43:03.094260 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6389-account-create-2pp9j"] Nov 24 13:43:05 crc kubenswrapper[4824]: I1124 13:43:05.023203 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ac2f711-f52c-4424-930b-a40967d136c6" path="/var/lib/kubelet/pods/5ac2f711-f52c-4424-930b-a40967d136c6/volumes" Nov 24 13:43:05 crc kubenswrapper[4824]: I1124 13:43:05.025269 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6faec986-3929-49c0-98e4-021352f2a8b9" path="/var/lib/kubelet/pods/6faec986-3929-49c0-98e4-021352f2a8b9/volumes" Nov 24 13:43:08 crc kubenswrapper[4824]: I1124 13:43:08.010200 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:43:08 crc kubenswrapper[4824]: E1124 13:43:08.010485 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:43:21 crc kubenswrapper[4824]: I1124 13:43:21.010660 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:43:21 crc kubenswrapper[4824]: E1124 13:43:21.011674 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:43:22 crc kubenswrapper[4824]: I1124 13:43:22.329203 4824 scope.go:117] "RemoveContainer" containerID="75c44665ccbe357a4934e151d2bca95543425285f24204ea3fd2054a8d3a2cec" Nov 24 13:43:22 crc kubenswrapper[4824]: I1124 13:43:22.350110 4824 scope.go:117] "RemoveContainer" containerID="a3199cceb1a1590bb1f9f6496de156eed1e565bfdb2b9afd81cb3775dbe9f5ec" Nov 24 13:43:22 crc kubenswrapper[4824]: I1124 13:43:22.374418 4824 scope.go:117] "RemoveContainer" containerID="fc12f9e41717cf3b6f976eb4a1b63c30c565fc37de6118c68070f24a1120cee2" Nov 24 13:43:22 crc kubenswrapper[4824]: I1124 13:43:22.429582 4824 scope.go:117] "RemoveContainer" containerID="7fc1080cbb820bcb5d38751e729538916b95a8ba3d604c89c2eaf90a17f24b73" Nov 24 13:43:22 crc kubenswrapper[4824]: I1124 13:43:22.475737 4824 scope.go:117] "RemoveContainer" containerID="14383d0ba2a9e4d43349c9d7a0c19a2023268e801a841a423c5134f7f69ed804" Nov 24 13:43:22 crc kubenswrapper[4824]: I1124 13:43:22.510465 4824 scope.go:117] "RemoveContainer" containerID="987c50051bd0e9f80f7397d64a85e667b6a2c95451fc31064e48afcda69b206b" Nov 24 13:43:22 crc kubenswrapper[4824]: I1124 13:43:22.552567 4824 scope.go:117] "RemoveContainer" containerID="756cabcd5fa409c5ef99a5b9bd605d437bf5427e3e473f4d97178ff72772029a" Nov 24 13:43:23 crc kubenswrapper[4824]: I1124 13:43:23.917703 4824 generic.go:334] "Generic (PLEG): container finished" podID="5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7" containerID="dba1dac792c06b61f5bee5f4fcb20f3fcea0c9403c2ba4da3eb01693676fd5c5" exitCode=0 Nov 24 13:43:23 crc kubenswrapper[4824]: I1124 13:43:23.917774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" event={"ID":"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7","Type":"ContainerDied","Data":"dba1dac792c06b61f5bee5f4fcb20f3fcea0c9403c2ba4da3eb01693676fd5c5"} Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.389629 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.521507 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-inventory\") pod \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.521646 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-ssh-key\") pod \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.521725 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-bootstrap-combined-ca-bundle\") pod \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.521861 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2vf5\" (UniqueName: \"kubernetes.io/projected/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-kube-api-access-m2vf5\") pod \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\" (UID: \"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7\") " Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.527672 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-kube-api-access-m2vf5" (OuterVolumeSpecName: "kube-api-access-m2vf5") pod "5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7" (UID: "5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7"). InnerVolumeSpecName "kube-api-access-m2vf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.527905 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7" (UID: "5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.554111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-inventory" (OuterVolumeSpecName: "inventory") pod "5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7" (UID: "5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.554637 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7" (UID: "5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.624414 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.624446 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2vf5\" (UniqueName: \"kubernetes.io/projected/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-kube-api-access-m2vf5\") on node \"crc\" DevicePath \"\"" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.624459 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.624467 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.942621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" event={"ID":"5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7","Type":"ContainerDied","Data":"7e5e8bbf0e27e6560abe1211e6efea39f88dd6a3cb96c49120d5eb77bfbcdbb9"} Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.942669 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e5e8bbf0e27e6560abe1211e6efea39f88dd6a3cb96c49120d5eb77bfbcdbb9" Nov 24 13:43:25 crc kubenswrapper[4824]: I1124 13:43:25.942704 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.031666 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm"] Nov 24 13:43:26 crc kubenswrapper[4824]: E1124 13:43:26.032036 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.032057 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 13:43:26 crc kubenswrapper[4824]: E1124 13:43:26.032103 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerName="extract-content" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.032111 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerName="extract-content" Nov 24 13:43:26 crc kubenswrapper[4824]: E1124 13:43:26.032120 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerName="extract-utilities" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.032126 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerName="extract-utilities" Nov 24 13:43:26 crc kubenswrapper[4824]: E1124 13:43:26.032134 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerName="registry-server" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.032139 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerName="registry-server" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.032295 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="dec682c6-9c93-4e5f-a595-51a31d308c46" containerName="registry-server" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.032327 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.032907 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.037618 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.037723 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.037866 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.039126 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.066153 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm"] Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.133013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.133187 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.133278 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c25z6\" (UniqueName: \"kubernetes.io/projected/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-kube-api-access-c25z6\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.235432 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.235738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.235840 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c25z6\" (UniqueName: \"kubernetes.io/projected/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-kube-api-access-c25z6\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.242595 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.244285 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.257706 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c25z6\" (UniqueName: \"kubernetes.io/projected/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-kube-api-access-c25z6\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.355434 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.863165 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm"] Nov 24 13:43:26 crc kubenswrapper[4824]: I1124 13:43:26.962342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" event={"ID":"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada","Type":"ContainerStarted","Data":"248ab970d3dc9cb975e8544b4908e0e1f3ce6ba573d4bdb0ecaa6a03115cbe32"} Nov 24 13:43:27 crc kubenswrapper[4824]: I1124 13:43:27.972447 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" event={"ID":"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada","Type":"ContainerStarted","Data":"b523eb59802a5d93e71efcd9b858fa23c931a9bec161f937f1098ec0c4f4faf8"} Nov 24 13:43:27 crc kubenswrapper[4824]: I1124 13:43:27.991913 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" podStartSLOduration=1.8132042830000001 podStartE2EDuration="1.9917829s" podCreationTimestamp="2025-11-24 13:43:26 +0000 UTC" firstStartedPulling="2025-11-24 13:43:26.867612142 +0000 UTC m=+1748.507151452" lastFinishedPulling="2025-11-24 13:43:27.046190759 +0000 UTC m=+1748.685730069" observedRunningTime="2025-11-24 13:43:27.9913785 +0000 UTC m=+1749.630917820" watchObservedRunningTime="2025-11-24 13:43:27.9917829 +0000 UTC m=+1749.631322210" Nov 24 13:43:28 crc kubenswrapper[4824]: I1124 13:43:28.039779 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-76b8-account-create-945pv"] Nov 24 13:43:28 crc kubenswrapper[4824]: I1124 13:43:28.047265 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-76b8-account-create-945pv"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.023775 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="945888ec-9f85-407e-b7d1-1feaac5c5e48" path="/var/lib/kubelet/pods/945888ec-9f85-407e-b7d1-1feaac5c5e48/volumes" Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.057981 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-ltp6q"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.065366 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-829a-account-create-zt6bp"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.073255 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-nzwdl"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.080227 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-lp9lk"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.087694 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-4b88-account-create-ld6br"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.095645 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-lp9lk"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.103195 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-4b88-account-create-ld6br"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.109902 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-829a-account-create-zt6bp"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.116455 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-nzwdl"] Nov 24 13:43:29 crc kubenswrapper[4824]: I1124 13:43:29.123431 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-ltp6q"] Nov 24 13:43:31 crc kubenswrapper[4824]: I1124 13:43:31.023251 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38a48af4-2b60-429d-9037-e483a357cde3" path="/var/lib/kubelet/pods/38a48af4-2b60-429d-9037-e483a357cde3/volumes" Nov 24 13:43:31 crc kubenswrapper[4824]: I1124 13:43:31.025725 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c0a2dfb-5b32-409d-9f07-87af1957f886" path="/var/lib/kubelet/pods/3c0a2dfb-5b32-409d-9f07-87af1957f886/volumes" Nov 24 13:43:31 crc kubenswrapper[4824]: I1124 13:43:31.026903 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f77f534-a41d-46dc-9ac6-faf212fed300" path="/var/lib/kubelet/pods/6f77f534-a41d-46dc-9ac6-faf212fed300/volumes" Nov 24 13:43:31 crc kubenswrapper[4824]: I1124 13:43:31.027794 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77e59dd3-1218-4204-8573-24c3fa35d41b" path="/var/lib/kubelet/pods/77e59dd3-1218-4204-8573-24c3fa35d41b/volumes" Nov 24 13:43:31 crc kubenswrapper[4824]: I1124 13:43:31.029405 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af46b974-486a-4b7c-8102-1c70f023cac2" path="/var/lib/kubelet/pods/af46b974-486a-4b7c-8102-1c70f023cac2/volumes" Nov 24 13:43:33 crc kubenswrapper[4824]: I1124 13:43:33.011148 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:43:33 crc kubenswrapper[4824]: E1124 13:43:33.012082 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:43:34 crc kubenswrapper[4824]: I1124 13:43:34.038248 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-x4w8p"] Nov 24 13:43:34 crc kubenswrapper[4824]: I1124 13:43:34.063852 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-x4w8p"] Nov 24 13:43:35 crc kubenswrapper[4824]: I1124 13:43:35.031955 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13797045-20bc-4f20-bf13-47b71e2f3342" path="/var/lib/kubelet/pods/13797045-20bc-4f20-bf13-47b71e2f3342/volumes" Nov 24 13:43:36 crc kubenswrapper[4824]: I1124 13:43:36.031580 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-sb5fc"] Nov 24 13:43:36 crc kubenswrapper[4824]: I1124 13:43:36.042977 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-sb5fc"] Nov 24 13:43:37 crc kubenswrapper[4824]: I1124 13:43:37.020164 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b623f0c-385a-465d-9132-9406dfdec4ec" path="/var/lib/kubelet/pods/0b623f0c-385a-465d-9132-9406dfdec4ec/volumes" Nov 24 13:43:46 crc kubenswrapper[4824]: I1124 13:43:46.010576 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:43:46 crc kubenswrapper[4824]: E1124 13:43:46.011463 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:44:01 crc kubenswrapper[4824]: I1124 13:44:01.011614 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:44:01 crc kubenswrapper[4824]: E1124 13:44:01.012681 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:44:15 crc kubenswrapper[4824]: I1124 13:44:15.057137 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-95tmx"] Nov 24 13:44:15 crc kubenswrapper[4824]: I1124 13:44:15.075828 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-95tmx"] Nov 24 13:44:16 crc kubenswrapper[4824]: I1124 13:44:16.011871 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:44:16 crc kubenswrapper[4824]: E1124 13:44:16.012426 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:44:17 crc kubenswrapper[4824]: I1124 13:44:17.022927 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e88a010-ca80-44fc-bd45-d741e3acdc47" path="/var/lib/kubelet/pods/8e88a010-ca80-44fc-bd45-d741e3acdc47/volumes" Nov 24 13:44:22 crc kubenswrapper[4824]: I1124 13:44:22.750390 4824 scope.go:117] "RemoveContainer" containerID="23c2b8f40274336e54fad1534e72aeb73219e9b1b9067768f4767000eea23a50" Nov 24 13:44:22 crc kubenswrapper[4824]: I1124 13:44:22.783386 4824 scope.go:117] "RemoveContainer" containerID="f4730305e6c94d84d1983d5c4b44d3b9ba745e9575bf6bad94a7cd36271308f9" Nov 24 13:44:22 crc kubenswrapper[4824]: I1124 13:44:22.835898 4824 scope.go:117] "RemoveContainer" containerID="4fe10b3cae19503781c7eea75a17365051a790c7b797f76de1ca4c3a7db834a3" Nov 24 13:44:22 crc kubenswrapper[4824]: I1124 13:44:22.896998 4824 scope.go:117] "RemoveContainer" containerID="5bd31d9784253bd14a5bdc2e681826c9955f8928e1df547da0f6faee12c815dc" Nov 24 13:44:22 crc kubenswrapper[4824]: I1124 13:44:22.926058 4824 scope.go:117] "RemoveContainer" containerID="a0f8b224e94e354de91ffe9c2ed72b311c9afd81a58cc6d65e1f257ea59e8c42" Nov 24 13:44:22 crc kubenswrapper[4824]: I1124 13:44:22.967714 4824 scope.go:117] "RemoveContainer" containerID="46993937eaf72dd5104a840a6466f2d6efeddd43a101453f90366c039138f1b9" Nov 24 13:44:23 crc kubenswrapper[4824]: I1124 13:44:23.013144 4824 scope.go:117] "RemoveContainer" containerID="9c615a9cf2ff228c5f8bbb889943d717da8081ca0eaae2ba5b6be1b6cb28445a" Nov 24 13:44:23 crc kubenswrapper[4824]: I1124 13:44:23.033308 4824 scope.go:117] "RemoveContainer" containerID="d1f53bd96f5a5dcc7e5c904637aeed92816ecb28bbb1dfef4a00c402c27794df" Nov 24 13:44:23 crc kubenswrapper[4824]: I1124 13:44:23.054774 4824 scope.go:117] "RemoveContainer" containerID="3235005020b9c2b3de53baed0de55b2479d2b7cb39773877fd34bc2c12a625ef" Nov 24 13:44:27 crc kubenswrapper[4824]: I1124 13:44:27.034283 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nhzcv"] Nov 24 13:44:27 crc kubenswrapper[4824]: I1124 13:44:27.043195 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nhzcv"] Nov 24 13:44:28 crc kubenswrapper[4824]: I1124 13:44:28.010620 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:44:28 crc kubenswrapper[4824]: E1124 13:44:28.011329 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:44:28 crc kubenswrapper[4824]: I1124 13:44:28.039107 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-8wqh7"] Nov 24 13:44:28 crc kubenswrapper[4824]: I1124 13:44:28.048204 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-trpzx"] Nov 24 13:44:28 crc kubenswrapper[4824]: I1124 13:44:28.058316 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-8wqh7"] Nov 24 13:44:28 crc kubenswrapper[4824]: I1124 13:44:28.066520 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-trpzx"] Nov 24 13:44:29 crc kubenswrapper[4824]: I1124 13:44:29.024638 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0781aef7-bf19-478e-931f-a10a3dcd586d" path="/var/lib/kubelet/pods/0781aef7-bf19-478e-931f-a10a3dcd586d/volumes" Nov 24 13:44:29 crc kubenswrapper[4824]: I1124 13:44:29.025690 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4735a096-fa3b-4a42-8ff1-fb1f92b2e953" path="/var/lib/kubelet/pods/4735a096-fa3b-4a42-8ff1-fb1f92b2e953/volumes" Nov 24 13:44:29 crc kubenswrapper[4824]: I1124 13:44:29.026616 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef6bc1b-dfa9-4c16-9940-420f9a7a43ce" path="/var/lib/kubelet/pods/aef6bc1b-dfa9-4c16-9940-420f9a7a43ce/volumes" Nov 24 13:44:43 crc kubenswrapper[4824]: I1124 13:44:43.012893 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:44:43 crc kubenswrapper[4824]: E1124 13:44:43.013894 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:44:45 crc kubenswrapper[4824]: I1124 13:44:45.047916 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-69zcz"] Nov 24 13:44:45 crc kubenswrapper[4824]: I1124 13:44:45.060552 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-69zcz"] Nov 24 13:44:47 crc kubenswrapper[4824]: I1124 13:44:47.028007 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1af3874-3566-4785-a5fe-13375aa658c3" path="/var/lib/kubelet/pods/b1af3874-3566-4785-a5fe-13375aa658c3/volumes" Nov 24 13:44:57 crc kubenswrapper[4824]: I1124 13:44:57.010461 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:44:57 crc kubenswrapper[4824]: E1124 13:44:57.011353 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.151730 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj"] Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.153629 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.158395 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.158709 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.158899 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj"] Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.319165 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f59d69e2-3080-431f-8b9b-c775a6397109-config-volume\") pod \"collect-profiles-29399865-x95hj\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.319460 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f59d69e2-3080-431f-8b9b-c775a6397109-secret-volume\") pod \"collect-profiles-29399865-x95hj\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.319605 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tdvz\" (UniqueName: \"kubernetes.io/projected/f59d69e2-3080-431f-8b9b-c775a6397109-kube-api-access-8tdvz\") pod \"collect-profiles-29399865-x95hj\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.421962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f59d69e2-3080-431f-8b9b-c775a6397109-config-volume\") pod \"collect-profiles-29399865-x95hj\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.422022 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f59d69e2-3080-431f-8b9b-c775a6397109-secret-volume\") pod \"collect-profiles-29399865-x95hj\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.422087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tdvz\" (UniqueName: \"kubernetes.io/projected/f59d69e2-3080-431f-8b9b-c775a6397109-kube-api-access-8tdvz\") pod \"collect-profiles-29399865-x95hj\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.422985 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f59d69e2-3080-431f-8b9b-c775a6397109-config-volume\") pod \"collect-profiles-29399865-x95hj\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.430408 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f59d69e2-3080-431f-8b9b-c775a6397109-secret-volume\") pod \"collect-profiles-29399865-x95hj\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.443849 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tdvz\" (UniqueName: \"kubernetes.io/projected/f59d69e2-3080-431f-8b9b-c775a6397109-kube-api-access-8tdvz\") pod \"collect-profiles-29399865-x95hj\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.472098 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:00 crc kubenswrapper[4824]: I1124 13:45:00.963284 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj"] Nov 24 13:45:01 crc kubenswrapper[4824]: I1124 13:45:01.882273 4824 generic.go:334] "Generic (PLEG): container finished" podID="f59d69e2-3080-431f-8b9b-c775a6397109" containerID="d1f7b025fb9f3898fbe0d0c327b1877f2a52caca91c9e7f7990ae00f9b428d55" exitCode=0 Nov 24 13:45:01 crc kubenswrapper[4824]: I1124 13:45:01.882573 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" event={"ID":"f59d69e2-3080-431f-8b9b-c775a6397109","Type":"ContainerDied","Data":"d1f7b025fb9f3898fbe0d0c327b1877f2a52caca91c9e7f7990ae00f9b428d55"} Nov 24 13:45:01 crc kubenswrapper[4824]: I1124 13:45:01.883656 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" event={"ID":"f59d69e2-3080-431f-8b9b-c775a6397109","Type":"ContainerStarted","Data":"bcc04a4d96e7ebd19fbff3c1075e67eadf7eddadb195f4970c53e263b7c95c92"} Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.328955 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.475083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f59d69e2-3080-431f-8b9b-c775a6397109-config-volume\") pod \"f59d69e2-3080-431f-8b9b-c775a6397109\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.475190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdvz\" (UniqueName: \"kubernetes.io/projected/f59d69e2-3080-431f-8b9b-c775a6397109-kube-api-access-8tdvz\") pod \"f59d69e2-3080-431f-8b9b-c775a6397109\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.475217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f59d69e2-3080-431f-8b9b-c775a6397109-secret-volume\") pod \"f59d69e2-3080-431f-8b9b-c775a6397109\" (UID: \"f59d69e2-3080-431f-8b9b-c775a6397109\") " Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.476059 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f59d69e2-3080-431f-8b9b-c775a6397109-config-volume" (OuterVolumeSpecName: "config-volume") pod "f59d69e2-3080-431f-8b9b-c775a6397109" (UID: "f59d69e2-3080-431f-8b9b-c775a6397109"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.497030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f59d69e2-3080-431f-8b9b-c775a6397109-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f59d69e2-3080-431f-8b9b-c775a6397109" (UID: "f59d69e2-3080-431f-8b9b-c775a6397109"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.497034 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f59d69e2-3080-431f-8b9b-c775a6397109-kube-api-access-8tdvz" (OuterVolumeSpecName: "kube-api-access-8tdvz") pod "f59d69e2-3080-431f-8b9b-c775a6397109" (UID: "f59d69e2-3080-431f-8b9b-c775a6397109"). InnerVolumeSpecName "kube-api-access-8tdvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.576888 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdvz\" (UniqueName: \"kubernetes.io/projected/f59d69e2-3080-431f-8b9b-c775a6397109-kube-api-access-8tdvz\") on node \"crc\" DevicePath \"\"" Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.576932 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f59d69e2-3080-431f-8b9b-c775a6397109-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.576942 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f59d69e2-3080-431f-8b9b-c775a6397109-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.903735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" event={"ID":"f59d69e2-3080-431f-8b9b-c775a6397109","Type":"ContainerDied","Data":"bcc04a4d96e7ebd19fbff3c1075e67eadf7eddadb195f4970c53e263b7c95c92"} Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.903789 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcc04a4d96e7ebd19fbff3c1075e67eadf7eddadb195f4970c53e263b7c95c92" Nov 24 13:45:03 crc kubenswrapper[4824]: I1124 13:45:03.904193 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj" Nov 24 13:45:09 crc kubenswrapper[4824]: I1124 13:45:09.015980 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:45:09 crc kubenswrapper[4824]: E1124 13:45:09.016902 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:45:11 crc kubenswrapper[4824]: I1124 13:45:11.983597 4824 generic.go:334] "Generic (PLEG): container finished" podID="73fc4178-cf60-40cb-b3f0-ea7dc1b1bada" containerID="b523eb59802a5d93e71efcd9b858fa23c931a9bec161f937f1098ec0c4f4faf8" exitCode=0 Nov 24 13:45:11 crc kubenswrapper[4824]: I1124 13:45:11.983654 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" event={"ID":"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada","Type":"ContainerDied","Data":"b523eb59802a5d93e71efcd9b858fa23c931a9bec161f937f1098ec0c4f4faf8"} Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.386762 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.579157 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-inventory\") pod \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.579604 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-ssh-key\") pod \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.579725 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c25z6\" (UniqueName: \"kubernetes.io/projected/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-kube-api-access-c25z6\") pod \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\" (UID: \"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada\") " Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.598585 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-kube-api-access-c25z6" (OuterVolumeSpecName: "kube-api-access-c25z6") pod "73fc4178-cf60-40cb-b3f0-ea7dc1b1bada" (UID: "73fc4178-cf60-40cb-b3f0-ea7dc1b1bada"). InnerVolumeSpecName "kube-api-access-c25z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.627235 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-inventory" (OuterVolumeSpecName: "inventory") pod "73fc4178-cf60-40cb-b3f0-ea7dc1b1bada" (UID: "73fc4178-cf60-40cb-b3f0-ea7dc1b1bada"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.633641 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "73fc4178-cf60-40cb-b3f0-ea7dc1b1bada" (UID: "73fc4178-cf60-40cb-b3f0-ea7dc1b1bada"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.688058 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.688093 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:45:13 crc kubenswrapper[4824]: I1124 13:45:13.688105 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c25z6\" (UniqueName: \"kubernetes.io/projected/73fc4178-cf60-40cb-b3f0-ea7dc1b1bada-kube-api-access-c25z6\") on node \"crc\" DevicePath \"\"" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.008115 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" event={"ID":"73fc4178-cf60-40cb-b3f0-ea7dc1b1bada","Type":"ContainerDied","Data":"248ab970d3dc9cb975e8544b4908e0e1f3ce6ba573d4bdb0ecaa6a03115cbe32"} Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.008530 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="248ab970d3dc9cb975e8544b4908e0e1f3ce6ba573d4bdb0ecaa6a03115cbe32" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.008212 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.120706 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds"] Nov 24 13:45:14 crc kubenswrapper[4824]: E1124 13:45:14.121188 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f59d69e2-3080-431f-8b9b-c775a6397109" containerName="collect-profiles" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.121233 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f59d69e2-3080-431f-8b9b-c775a6397109" containerName="collect-profiles" Nov 24 13:45:14 crc kubenswrapper[4824]: E1124 13:45:14.121277 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73fc4178-cf60-40cb-b3f0-ea7dc1b1bada" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.121286 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="73fc4178-cf60-40cb-b3f0-ea7dc1b1bada" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.121525 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f59d69e2-3080-431f-8b9b-c775a6397109" containerName="collect-profiles" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.121554 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="73fc4178-cf60-40cb-b3f0-ea7dc1b1bada" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.122392 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.135133 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.135188 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.135374 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.135596 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.136287 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds"] Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.300122 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm7qw\" (UniqueName: \"kubernetes.io/projected/18493374-ba71-4ac2-97dd-171111137b81-kube-api-access-pm7qw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k5rds\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.300220 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k5rds\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.300342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k5rds\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.402084 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k5rds\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.402139 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm7qw\" (UniqueName: \"kubernetes.io/projected/18493374-ba71-4ac2-97dd-171111137b81-kube-api-access-pm7qw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k5rds\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.402194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k5rds\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.407600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k5rds\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.407911 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k5rds\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.421533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm7qw\" (UniqueName: \"kubernetes.io/projected/18493374-ba71-4ac2-97dd-171111137b81-kube-api-access-pm7qw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k5rds\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.439344 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:45:14 crc kubenswrapper[4824]: I1124 13:45:14.971578 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds"] Nov 24 13:45:15 crc kubenswrapper[4824]: I1124 13:45:15.024200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" event={"ID":"18493374-ba71-4ac2-97dd-171111137b81","Type":"ContainerStarted","Data":"c58de976856d642867c99478e228447f295a02df5688d7ae501adf6003d6253e"} Nov 24 13:45:16 crc kubenswrapper[4824]: I1124 13:45:16.034611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" event={"ID":"18493374-ba71-4ac2-97dd-171111137b81","Type":"ContainerStarted","Data":"e3ca70d033bd589d917e15b2fdfd5a29599d184edcd361a1ed6698ca8c6fe84c"} Nov 24 13:45:16 crc kubenswrapper[4824]: I1124 13:45:16.060105 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" podStartSLOduration=1.876654658 podStartE2EDuration="2.060084134s" podCreationTimestamp="2025-11-24 13:45:14 +0000 UTC" firstStartedPulling="2025-11-24 13:45:14.978965785 +0000 UTC m=+1856.618505095" lastFinishedPulling="2025-11-24 13:45:15.162395251 +0000 UTC m=+1856.801934571" observedRunningTime="2025-11-24 13:45:16.049490895 +0000 UTC m=+1857.689030225" watchObservedRunningTime="2025-11-24 13:45:16.060084134 +0000 UTC m=+1857.699623464" Nov 24 13:45:20 crc kubenswrapper[4824]: I1124 13:45:20.010134 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:45:20 crc kubenswrapper[4824]: E1124 13:45:20.010843 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:45:23 crc kubenswrapper[4824]: I1124 13:45:23.243829 4824 scope.go:117] "RemoveContainer" containerID="9362b01e70cca8836c8dea1c873faa494ce0ceebc8d74d961b4000e0192cc1b2" Nov 24 13:45:23 crc kubenswrapper[4824]: I1124 13:45:23.274867 4824 scope.go:117] "RemoveContainer" containerID="6761b3024842943636b1639e6b9bc2ce39c1fa754d686e618def474225da2dd0" Nov 24 13:45:23 crc kubenswrapper[4824]: I1124 13:45:23.355257 4824 scope.go:117] "RemoveContainer" containerID="0b569fad882da65e1f89b01252135d1af8a4c1c48dc1af5f9ec689f8f7dee040" Nov 24 13:45:23 crc kubenswrapper[4824]: I1124 13:45:23.400066 4824 scope.go:117] "RemoveContainer" containerID="e05c7e2a822f0ecac3863cc30b8a66eea207ed911e30a1572205c34f80baddfa" Nov 24 13:45:23 crc kubenswrapper[4824]: I1124 13:45:23.452274 4824 scope.go:117] "RemoveContainer" containerID="bbec7cff655730f600e0aa2221054da899c4997d2ea1937dd73b610767fa5b2c" Nov 24 13:45:23 crc kubenswrapper[4824]: I1124 13:45:23.481829 4824 scope.go:117] "RemoveContainer" containerID="434c07b273fa8d0de8a9e5b3e6e54b13b06274c92268cd16f763a58c828cb842" Nov 24 13:45:23 crc kubenswrapper[4824]: I1124 13:45:23.503621 4824 scope.go:117] "RemoveContainer" containerID="96d6e74afc275b4335f284c28a31f9d22e683851773e95ed6c8b4c1090eda72d" Nov 24 13:45:33 crc kubenswrapper[4824]: I1124 13:45:33.010937 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:45:33 crc kubenswrapper[4824]: E1124 13:45:33.012198 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.063007 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-5h6sl"] Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.080694 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-8dxxk"] Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.090065 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-mnhn5"] Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.104118 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-5h6sl"] Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.111181 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-8dxxk"] Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.117689 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0a19-account-create-gxxbb"] Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.123845 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f0d5-account-create-46mmw"] Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.129883 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f0d5-account-create-46mmw"] Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.138074 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0a19-account-create-gxxbb"] Nov 24 13:45:43 crc kubenswrapper[4824]: I1124 13:45:43.145465 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-mnhn5"] Nov 24 13:45:44 crc kubenswrapper[4824]: I1124 13:45:44.040609 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-e59f-account-create-t5q2j"] Nov 24 13:45:44 crc kubenswrapper[4824]: I1124 13:45:44.047275 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-e59f-account-create-t5q2j"] Nov 24 13:45:45 crc kubenswrapper[4824]: I1124 13:45:45.022451 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="570330a6-c219-4a73-8731-b85ca896c82c" path="/var/lib/kubelet/pods/570330a6-c219-4a73-8731-b85ca896c82c/volumes" Nov 24 13:45:45 crc kubenswrapper[4824]: I1124 13:45:45.023554 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cbf4587-2c54-4220-a514-62b043882b5b" path="/var/lib/kubelet/pods/6cbf4587-2c54-4220-a514-62b043882b5b/volumes" Nov 24 13:45:45 crc kubenswrapper[4824]: I1124 13:45:45.025392 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9dd708-d69f-4c23-bbe1-2f41c7f18fae" path="/var/lib/kubelet/pods/7e9dd708-d69f-4c23-bbe1-2f41c7f18fae/volumes" Nov 24 13:45:45 crc kubenswrapper[4824]: I1124 13:45:45.026459 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fc09e1d-e018-4086-b17e-f7a1ccf02951" path="/var/lib/kubelet/pods/8fc09e1d-e018-4086-b17e-f7a1ccf02951/volumes" Nov 24 13:45:45 crc kubenswrapper[4824]: I1124 13:45:45.028162 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f2ba678-6efb-4ec5-b163-da752d2edecb" path="/var/lib/kubelet/pods/9f2ba678-6efb-4ec5-b163-da752d2edecb/volumes" Nov 24 13:45:45 crc kubenswrapper[4824]: I1124 13:45:45.029434 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f396378e-6a0d-4853-a622-3d5ca30d45fc" path="/var/lib/kubelet/pods/f396378e-6a0d-4853-a622-3d5ca30d45fc/volumes" Nov 24 13:45:48 crc kubenswrapper[4824]: I1124 13:45:48.011101 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:45:48 crc kubenswrapper[4824]: E1124 13:45:48.011774 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:46:03 crc kubenswrapper[4824]: I1124 13:46:03.010824 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:46:03 crc kubenswrapper[4824]: E1124 13:46:03.011616 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:46:14 crc kubenswrapper[4824]: I1124 13:46:14.042609 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7xbd2"] Nov 24 13:46:14 crc kubenswrapper[4824]: I1124 13:46:14.056179 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7xbd2"] Nov 24 13:46:15 crc kubenswrapper[4824]: I1124 13:46:15.020731 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a86b9f-f865-4d11-8d95-ac103ea18ca7" path="/var/lib/kubelet/pods/83a86b9f-f865-4d11-8d95-ac103ea18ca7/volumes" Nov 24 13:46:18 crc kubenswrapper[4824]: I1124 13:46:18.010877 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:46:18 crc kubenswrapper[4824]: E1124 13:46:18.012609 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:46:23 crc kubenswrapper[4824]: I1124 13:46:23.652952 4824 scope.go:117] "RemoveContainer" containerID="b73f7af5aa4e5276937e5032ce17b43502984db3fb1f39f12e321fca6cfd499c" Nov 24 13:46:23 crc kubenswrapper[4824]: I1124 13:46:23.687003 4824 scope.go:117] "RemoveContainer" containerID="f1d7def3cc42ff723e398a77610016b05a96420dd0b60e3e0395650b2a042e20" Nov 24 13:46:23 crc kubenswrapper[4824]: I1124 13:46:23.737151 4824 scope.go:117] "RemoveContainer" containerID="54da7f04b24314fc7fa59d93fb6b64cceea5b86eb0b61c31358ef6fc00767738" Nov 24 13:46:23 crc kubenswrapper[4824]: I1124 13:46:23.775065 4824 scope.go:117] "RemoveContainer" containerID="83ad15175826da0d74b1ff9e2c38c2c74235e4e15b71addb9712beac243cd8ca" Nov 24 13:46:23 crc kubenswrapper[4824]: I1124 13:46:23.843458 4824 scope.go:117] "RemoveContainer" containerID="096465ccc50d35cad02903956880aa23ed4855202d5dbbf17089b3b383417e33" Nov 24 13:46:23 crc kubenswrapper[4824]: I1124 13:46:23.873545 4824 scope.go:117] "RemoveContainer" containerID="4252cad99b13bf30db5f9737bcbc367232d45b66ae8feb39f1ed0f5af8ef92eb" Nov 24 13:46:23 crc kubenswrapper[4824]: I1124 13:46:23.903906 4824 scope.go:117] "RemoveContainer" containerID="e7bd5a1cf8ef04bc77baede9980f996b19cd411f7bd846c1b620c067ba0cdee1" Nov 24 13:46:29 crc kubenswrapper[4824]: I1124 13:46:29.897764 4824 generic.go:334] "Generic (PLEG): container finished" podID="18493374-ba71-4ac2-97dd-171111137b81" containerID="e3ca70d033bd589d917e15b2fdfd5a29599d184edcd361a1ed6698ca8c6fe84c" exitCode=0 Nov 24 13:46:29 crc kubenswrapper[4824]: I1124 13:46:29.897967 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" event={"ID":"18493374-ba71-4ac2-97dd-171111137b81","Type":"ContainerDied","Data":"e3ca70d033bd589d917e15b2fdfd5a29599d184edcd361a1ed6698ca8c6fe84c"} Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.010910 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:46:31 crc kubenswrapper[4824]: E1124 13:46:31.011425 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.358001 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.456431 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-ssh-key\") pod \"18493374-ba71-4ac2-97dd-171111137b81\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.457063 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm7qw\" (UniqueName: \"kubernetes.io/projected/18493374-ba71-4ac2-97dd-171111137b81-kube-api-access-pm7qw\") pod \"18493374-ba71-4ac2-97dd-171111137b81\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.457226 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-inventory\") pod \"18493374-ba71-4ac2-97dd-171111137b81\" (UID: \"18493374-ba71-4ac2-97dd-171111137b81\") " Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.464348 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18493374-ba71-4ac2-97dd-171111137b81-kube-api-access-pm7qw" (OuterVolumeSpecName: "kube-api-access-pm7qw") pod "18493374-ba71-4ac2-97dd-171111137b81" (UID: "18493374-ba71-4ac2-97dd-171111137b81"). InnerVolumeSpecName "kube-api-access-pm7qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.484040 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18493374-ba71-4ac2-97dd-171111137b81" (UID: "18493374-ba71-4ac2-97dd-171111137b81"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.484558 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-inventory" (OuterVolumeSpecName: "inventory") pod "18493374-ba71-4ac2-97dd-171111137b81" (UID: "18493374-ba71-4ac2-97dd-171111137b81"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.560341 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.560377 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm7qw\" (UniqueName: \"kubernetes.io/projected/18493374-ba71-4ac2-97dd-171111137b81-kube-api-access-pm7qw\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.560389 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18493374-ba71-4ac2-97dd-171111137b81-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.932545 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" event={"ID":"18493374-ba71-4ac2-97dd-171111137b81","Type":"ContainerDied","Data":"c58de976856d642867c99478e228447f295a02df5688d7ae501adf6003d6253e"} Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.932611 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c58de976856d642867c99478e228447f295a02df5688d7ae501adf6003d6253e" Nov 24 13:46:31 crc kubenswrapper[4824]: I1124 13:46:31.932646 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k5rds" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.048736 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s"] Nov 24 13:46:32 crc kubenswrapper[4824]: E1124 13:46:32.049425 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18493374-ba71-4ac2-97dd-171111137b81" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.049449 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="18493374-ba71-4ac2-97dd-171111137b81" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.049784 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="18493374-ba71-4ac2-97dd-171111137b81" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.050968 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.065660 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.065919 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.066061 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.066382 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.079857 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s"] Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.176876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7nmw\" (UniqueName: \"kubernetes.io/projected/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-kube-api-access-t7nmw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.176926 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.177161 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.278970 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7nmw\" (UniqueName: \"kubernetes.io/projected/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-kube-api-access-t7nmw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.279035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.279117 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.284050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.286560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.311169 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7nmw\" (UniqueName: \"kubernetes.io/projected/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-kube-api-access-t7nmw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:32 crc kubenswrapper[4824]: I1124 13:46:32.384352 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:33 crc kubenswrapper[4824]: I1124 13:46:33.004894 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s"] Nov 24 13:46:33 crc kubenswrapper[4824]: I1124 13:46:33.951878 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" event={"ID":"ddb976c7-8261-4a6b-96a1-75a8e4b5d282","Type":"ContainerStarted","Data":"f60905d0033bbe294536271d76986d50d3e10fd090e09c931031cdaf9f43394a"} Nov 24 13:46:33 crc kubenswrapper[4824]: I1124 13:46:33.952283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" event={"ID":"ddb976c7-8261-4a6b-96a1-75a8e4b5d282","Type":"ContainerStarted","Data":"059181f0a935a5ec19bbe4033bd96a93a0ff932cbcece89b3433bf691a837e86"} Nov 24 13:46:33 crc kubenswrapper[4824]: I1124 13:46:33.979151 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" podStartSLOduration=1.82772428 podStartE2EDuration="1.979131332s" podCreationTimestamp="2025-11-24 13:46:32 +0000 UTC" firstStartedPulling="2025-11-24 13:46:33.005632689 +0000 UTC m=+1934.645172039" lastFinishedPulling="2025-11-24 13:46:33.157039781 +0000 UTC m=+1934.796579091" observedRunningTime="2025-11-24 13:46:33.968526962 +0000 UTC m=+1935.608066312" watchObservedRunningTime="2025-11-24 13:46:33.979131332 +0000 UTC m=+1935.618670652" Nov 24 13:46:39 crc kubenswrapper[4824]: I1124 13:46:39.001431 4824 generic.go:334] "Generic (PLEG): container finished" podID="ddb976c7-8261-4a6b-96a1-75a8e4b5d282" containerID="f60905d0033bbe294536271d76986d50d3e10fd090e09c931031cdaf9f43394a" exitCode=0 Nov 24 13:46:39 crc kubenswrapper[4824]: I1124 13:46:39.001573 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" event={"ID":"ddb976c7-8261-4a6b-96a1-75a8e4b5d282","Type":"ContainerDied","Data":"f60905d0033bbe294536271d76986d50d3e10fd090e09c931031cdaf9f43394a"} Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.541472 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.655821 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7nmw\" (UniqueName: \"kubernetes.io/projected/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-kube-api-access-t7nmw\") pod \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.655930 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-inventory\") pod \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.656066 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-ssh-key\") pod \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\" (UID: \"ddb976c7-8261-4a6b-96a1-75a8e4b5d282\") " Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.668403 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-kube-api-access-t7nmw" (OuterVolumeSpecName: "kube-api-access-t7nmw") pod "ddb976c7-8261-4a6b-96a1-75a8e4b5d282" (UID: "ddb976c7-8261-4a6b-96a1-75a8e4b5d282"). InnerVolumeSpecName "kube-api-access-t7nmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.690407 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-inventory" (OuterVolumeSpecName: "inventory") pod "ddb976c7-8261-4a6b-96a1-75a8e4b5d282" (UID: "ddb976c7-8261-4a6b-96a1-75a8e4b5d282"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.701693 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ddb976c7-8261-4a6b-96a1-75a8e4b5d282" (UID: "ddb976c7-8261-4a6b-96a1-75a8e4b5d282"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.762961 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.763003 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7nmw\" (UniqueName: \"kubernetes.io/projected/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-kube-api-access-t7nmw\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:40 crc kubenswrapper[4824]: I1124 13:46:40.763021 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddb976c7-8261-4a6b-96a1-75a8e4b5d282-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.028628 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.032262 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s" event={"ID":"ddb976c7-8261-4a6b-96a1-75a8e4b5d282","Type":"ContainerDied","Data":"059181f0a935a5ec19bbe4033bd96a93a0ff932cbcece89b3433bf691a837e86"} Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.032413 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="059181f0a935a5ec19bbe4033bd96a93a0ff932cbcece89b3433bf691a837e86" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.083218 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-9sr7j"] Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.107069 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-9sr7j"] Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.217196 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn"] Nov 24 13:46:41 crc kubenswrapper[4824]: E1124 13:46:41.218021 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb976c7-8261-4a6b-96a1-75a8e4b5d282" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.218043 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb976c7-8261-4a6b-96a1-75a8e4b5d282" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.218298 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddb976c7-8261-4a6b-96a1-75a8e4b5d282" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.219087 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.222372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.222739 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.222899 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.223013 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.227071 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn"] Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.271768 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pkhb\" (UniqueName: \"kubernetes.io/projected/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-kube-api-access-5pkhb\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nhzbn\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.271909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nhzbn\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.271988 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nhzbn\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.373950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pkhb\" (UniqueName: \"kubernetes.io/projected/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-kube-api-access-5pkhb\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nhzbn\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.374108 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nhzbn\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.374218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nhzbn\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.379100 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nhzbn\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.380176 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nhzbn\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.389053 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pkhb\" (UniqueName: \"kubernetes.io/projected/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-kube-api-access-5pkhb\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nhzbn\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:41 crc kubenswrapper[4824]: I1124 13:46:41.545946 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:46:42 crc kubenswrapper[4824]: I1124 13:46:42.036240 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hgk58"] Nov 24 13:46:42 crc kubenswrapper[4824]: I1124 13:46:42.047130 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hgk58"] Nov 24 13:46:42 crc kubenswrapper[4824]: I1124 13:46:42.169711 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn"] Nov 24 13:46:43 crc kubenswrapper[4824]: I1124 13:46:43.023389 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="001160fa-3d5e-4c31-bc99-d8cdd7b81324" path="/var/lib/kubelet/pods/001160fa-3d5e-4c31-bc99-d8cdd7b81324/volumes" Nov 24 13:46:43 crc kubenswrapper[4824]: I1124 13:46:43.024778 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e361ab6-07b2-49fb-ada0-80eff17ca67b" path="/var/lib/kubelet/pods/8e361ab6-07b2-49fb-ada0-80eff17ca67b/volumes" Nov 24 13:46:43 crc kubenswrapper[4824]: I1124 13:46:43.051078 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" event={"ID":"27ecdc8a-4391-4eb5-a900-0f28984fd1d4","Type":"ContainerStarted","Data":"63beee76dc4c05cef0a90027029729f87e63093f1d9005c9c65d6a32b85a4185"} Nov 24 13:46:43 crc kubenswrapper[4824]: I1124 13:46:43.051114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" event={"ID":"27ecdc8a-4391-4eb5-a900-0f28984fd1d4","Type":"ContainerStarted","Data":"a85e6d6f55b7d15a4d397a0cbb842e60f92ed8c07edde50b01e2ee6dd6184bc2"} Nov 24 13:46:43 crc kubenswrapper[4824]: I1124 13:46:43.079231 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" podStartSLOduration=1.91269198 podStartE2EDuration="2.079208441s" podCreationTimestamp="2025-11-24 13:46:41 +0000 UTC" firstStartedPulling="2025-11-24 13:46:42.159108398 +0000 UTC m=+1943.798647708" lastFinishedPulling="2025-11-24 13:46:42.325624819 +0000 UTC m=+1943.965164169" observedRunningTime="2025-11-24 13:46:43.06894328 +0000 UTC m=+1944.708482590" watchObservedRunningTime="2025-11-24 13:46:43.079208441 +0000 UTC m=+1944.718747751" Nov 24 13:46:44 crc kubenswrapper[4824]: I1124 13:46:44.010626 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:46:44 crc kubenswrapper[4824]: E1124 13:46:44.011593 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:46:59 crc kubenswrapper[4824]: I1124 13:46:59.015630 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:46:59 crc kubenswrapper[4824]: E1124 13:46:59.016448 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:47:14 crc kubenswrapper[4824]: I1124 13:47:14.010056 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:47:14 crc kubenswrapper[4824]: E1124 13:47:14.012531 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:47:24 crc kubenswrapper[4824]: I1124 13:47:24.052846 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-q9524"] Nov 24 13:47:24 crc kubenswrapper[4824]: I1124 13:47:24.065426 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-q9524"] Nov 24 13:47:24 crc kubenswrapper[4824]: I1124 13:47:24.072512 4824 scope.go:117] "RemoveContainer" containerID="5c5827b6aa081443eea4c5a8b80a7f5900c63746b3dbc07312ffbc7f66fde317" Nov 24 13:47:24 crc kubenswrapper[4824]: I1124 13:47:24.133132 4824 scope.go:117] "RemoveContainer" containerID="8e152f04adaf5d5c412c328288fc0c2b10cf5c58a2c777e18a964a59a7ff59e7" Nov 24 13:47:24 crc kubenswrapper[4824]: I1124 13:47:24.434945 4824 generic.go:334] "Generic (PLEG): container finished" podID="27ecdc8a-4391-4eb5-a900-0f28984fd1d4" containerID="63beee76dc4c05cef0a90027029729f87e63093f1d9005c9c65d6a32b85a4185" exitCode=0 Nov 24 13:47:24 crc kubenswrapper[4824]: I1124 13:47:24.435012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" event={"ID":"27ecdc8a-4391-4eb5-a900-0f28984fd1d4","Type":"ContainerDied","Data":"63beee76dc4c05cef0a90027029729f87e63093f1d9005c9c65d6a32b85a4185"} Nov 24 13:47:25 crc kubenswrapper[4824]: I1124 13:47:25.031471 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf" path="/var/lib/kubelet/pods/0c284e54-883c-4ea7-9ce2-bc6bc4a9c8cf/volumes" Nov 24 13:47:25 crc kubenswrapper[4824]: I1124 13:47:25.930557 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:47:25 crc kubenswrapper[4824]: I1124 13:47:25.983685 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-inventory\") pod \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " Nov 24 13:47:25 crc kubenswrapper[4824]: I1124 13:47:25.983780 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-ssh-key\") pod \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " Nov 24 13:47:25 crc kubenswrapper[4824]: I1124 13:47:25.983850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pkhb\" (UniqueName: \"kubernetes.io/projected/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-kube-api-access-5pkhb\") pod \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\" (UID: \"27ecdc8a-4391-4eb5-a900-0f28984fd1d4\") " Nov 24 13:47:25 crc kubenswrapper[4824]: I1124 13:47:25.997307 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-kube-api-access-5pkhb" (OuterVolumeSpecName: "kube-api-access-5pkhb") pod "27ecdc8a-4391-4eb5-a900-0f28984fd1d4" (UID: "27ecdc8a-4391-4eb5-a900-0f28984fd1d4"). InnerVolumeSpecName "kube-api-access-5pkhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.026915 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-inventory" (OuterVolumeSpecName: "inventory") pod "27ecdc8a-4391-4eb5-a900-0f28984fd1d4" (UID: "27ecdc8a-4391-4eb5-a900-0f28984fd1d4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.032604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27ecdc8a-4391-4eb5-a900-0f28984fd1d4" (UID: "27ecdc8a-4391-4eb5-a900-0f28984fd1d4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.086419 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.086450 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.086461 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pkhb\" (UniqueName: \"kubernetes.io/projected/27ecdc8a-4391-4eb5-a900-0f28984fd1d4-kube-api-access-5pkhb\") on node \"crc\" DevicePath \"\"" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.456827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" event={"ID":"27ecdc8a-4391-4eb5-a900-0f28984fd1d4","Type":"ContainerDied","Data":"a85e6d6f55b7d15a4d397a0cbb842e60f92ed8c07edde50b01e2ee6dd6184bc2"} Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.457176 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a85e6d6f55b7d15a4d397a0cbb842e60f92ed8c07edde50b01e2ee6dd6184bc2" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.456841 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nhzbn" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.581166 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8"] Nov 24 13:47:26 crc kubenswrapper[4824]: E1124 13:47:26.581694 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ecdc8a-4391-4eb5-a900-0f28984fd1d4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.581718 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ecdc8a-4391-4eb5-a900-0f28984fd1d4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.582019 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ecdc8a-4391-4eb5-a900-0f28984fd1d4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.582874 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.585193 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.585529 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.585652 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.587834 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.612450 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8"] Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.701387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.701785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.701934 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn474\" (UniqueName: \"kubernetes.io/projected/92ed5101-5b03-4be3-bd82-32ba9915ba15-kube-api-access-gn474\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.830863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.830971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.831091 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn474\" (UniqueName: \"kubernetes.io/projected/92ed5101-5b03-4be3-bd82-32ba9915ba15-kube-api-access-gn474\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.840526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.840605 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.852203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn474\" (UniqueName: \"kubernetes.io/projected/92ed5101-5b03-4be3-bd82-32ba9915ba15-kube-api-access-gn474\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:26 crc kubenswrapper[4824]: I1124 13:47:26.925893 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:47:27 crc kubenswrapper[4824]: I1124 13:47:27.468416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8"] Nov 24 13:47:28 crc kubenswrapper[4824]: I1124 13:47:28.010634 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:47:28 crc kubenswrapper[4824]: E1124 13:47:28.011548 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:47:28 crc kubenswrapper[4824]: I1124 13:47:28.481311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" event={"ID":"92ed5101-5b03-4be3-bd82-32ba9915ba15","Type":"ContainerStarted","Data":"339d32cac1461e4e1a6ffdeef7fb79d571e89f8553485706170f1f510770a8ea"} Nov 24 13:47:28 crc kubenswrapper[4824]: I1124 13:47:28.481619 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" event={"ID":"92ed5101-5b03-4be3-bd82-32ba9915ba15","Type":"ContainerStarted","Data":"f2fba5d92a271e0e96b88cc10b1ac35fe5950f8a3c9092d105de105847a3b0b8"} Nov 24 13:47:28 crc kubenswrapper[4824]: I1124 13:47:28.499614 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" podStartSLOduration=2.329148504 podStartE2EDuration="2.499593062s" podCreationTimestamp="2025-11-24 13:47:26 +0000 UTC" firstStartedPulling="2025-11-24 13:47:27.474322621 +0000 UTC m=+1989.113861931" lastFinishedPulling="2025-11-24 13:47:27.644767179 +0000 UTC m=+1989.284306489" observedRunningTime="2025-11-24 13:47:28.496044255 +0000 UTC m=+1990.135583565" watchObservedRunningTime="2025-11-24 13:47:28.499593062 +0000 UTC m=+1990.139132372" Nov 24 13:47:40 crc kubenswrapper[4824]: I1124 13:47:40.010745 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:47:40 crc kubenswrapper[4824]: E1124 13:47:40.011470 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:47:51 crc kubenswrapper[4824]: I1124 13:47:51.011130 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:47:51 crc kubenswrapper[4824]: I1124 13:47:51.741086 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"f7c32c7c1ab5c79ebcd1b5d0a45adaba8003211cdde83e67ef3cbc6580aa7fd2"} Nov 24 13:47:55 crc kubenswrapper[4824]: I1124 13:47:55.849481 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z58zc"] Nov 24 13:47:55 crc kubenswrapper[4824]: I1124 13:47:55.852492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:55 crc kubenswrapper[4824]: I1124 13:47:55.865561 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z58zc"] Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.046078 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b46ls\" (UniqueName: \"kubernetes.io/projected/d97159cf-d145-4107-8451-5ee2a8255079-kube-api-access-b46ls\") pod \"redhat-marketplace-z58zc\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.046144 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-catalog-content\") pod \"redhat-marketplace-z58zc\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.046418 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-utilities\") pod \"redhat-marketplace-z58zc\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.148467 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-catalog-content\") pod \"redhat-marketplace-z58zc\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.148583 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-utilities\") pod \"redhat-marketplace-z58zc\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.148703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b46ls\" (UniqueName: \"kubernetes.io/projected/d97159cf-d145-4107-8451-5ee2a8255079-kube-api-access-b46ls\") pod \"redhat-marketplace-z58zc\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.149410 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-catalog-content\") pod \"redhat-marketplace-z58zc\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.149596 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-utilities\") pod \"redhat-marketplace-z58zc\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.180156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b46ls\" (UniqueName: \"kubernetes.io/projected/d97159cf-d145-4107-8451-5ee2a8255079-kube-api-access-b46ls\") pod \"redhat-marketplace-z58zc\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.470450 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:47:56 crc kubenswrapper[4824]: I1124 13:47:56.976151 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z58zc"] Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.252087 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bdgcn"] Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.261828 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.265982 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdgcn"] Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.370126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-utilities\") pod \"redhat-operators-bdgcn\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.370198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg7b9\" (UniqueName: \"kubernetes.io/projected/974d2a77-e0fd-4a6d-8558-b8c4423777f6-kube-api-access-mg7b9\") pod \"redhat-operators-bdgcn\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.370262 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-catalog-content\") pod \"redhat-operators-bdgcn\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.472121 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-catalog-content\") pod \"redhat-operators-bdgcn\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.472290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-utilities\") pod \"redhat-operators-bdgcn\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.472358 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg7b9\" (UniqueName: \"kubernetes.io/projected/974d2a77-e0fd-4a6d-8558-b8c4423777f6-kube-api-access-mg7b9\") pod \"redhat-operators-bdgcn\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.472587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-catalog-content\") pod \"redhat-operators-bdgcn\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.472932 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-utilities\") pod \"redhat-operators-bdgcn\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.492194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg7b9\" (UniqueName: \"kubernetes.io/projected/974d2a77-e0fd-4a6d-8558-b8c4423777f6-kube-api-access-mg7b9\") pod \"redhat-operators-bdgcn\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.591568 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.792910 4824 generic.go:334] "Generic (PLEG): container finished" podID="d97159cf-d145-4107-8451-5ee2a8255079" containerID="a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941" exitCode=0 Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.793115 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z58zc" event={"ID":"d97159cf-d145-4107-8451-5ee2a8255079","Type":"ContainerDied","Data":"a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941"} Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.793138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z58zc" event={"ID":"d97159cf-d145-4107-8451-5ee2a8255079","Type":"ContainerStarted","Data":"20a26a86052dee32bae1745ce62d6ff566a57e5552847f654defe004cd32488a"} Nov 24 13:47:57 crc kubenswrapper[4824]: I1124 13:47:57.796311 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:47:58 crc kubenswrapper[4824]: I1124 13:47:58.065643 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdgcn"] Nov 24 13:47:58 crc kubenswrapper[4824]: W1124 13:47:58.075719 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod974d2a77_e0fd_4a6d_8558_b8c4423777f6.slice/crio-37b3b3a98f0a670a4a5919186963a92c724ff24b217acfd43a9ff8dfa39b706e WatchSource:0}: Error finding container 37b3b3a98f0a670a4a5919186963a92c724ff24b217acfd43a9ff8dfa39b706e: Status 404 returned error can't find the container with id 37b3b3a98f0a670a4a5919186963a92c724ff24b217acfd43a9ff8dfa39b706e Nov 24 13:47:58 crc kubenswrapper[4824]: I1124 13:47:58.808634 4824 generic.go:334] "Generic (PLEG): container finished" podID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerID="6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae" exitCode=0 Nov 24 13:47:58 crc kubenswrapper[4824]: I1124 13:47:58.808694 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdgcn" event={"ID":"974d2a77-e0fd-4a6d-8558-b8c4423777f6","Type":"ContainerDied","Data":"6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae"} Nov 24 13:47:58 crc kubenswrapper[4824]: I1124 13:47:58.808958 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdgcn" event={"ID":"974d2a77-e0fd-4a6d-8558-b8c4423777f6","Type":"ContainerStarted","Data":"37b3b3a98f0a670a4a5919186963a92c724ff24b217acfd43a9ff8dfa39b706e"} Nov 24 13:47:59 crc kubenswrapper[4824]: I1124 13:47:59.820008 4824 generic.go:334] "Generic (PLEG): container finished" podID="d97159cf-d145-4107-8451-5ee2a8255079" containerID="333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639" exitCode=0 Nov 24 13:47:59 crc kubenswrapper[4824]: I1124 13:47:59.820048 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z58zc" event={"ID":"d97159cf-d145-4107-8451-5ee2a8255079","Type":"ContainerDied","Data":"333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639"} Nov 24 13:48:00 crc kubenswrapper[4824]: I1124 13:48:00.829705 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z58zc" event={"ID":"d97159cf-d145-4107-8451-5ee2a8255079","Type":"ContainerStarted","Data":"68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3"} Nov 24 13:48:00 crc kubenswrapper[4824]: I1124 13:48:00.831228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdgcn" event={"ID":"974d2a77-e0fd-4a6d-8558-b8c4423777f6","Type":"ContainerStarted","Data":"16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077"} Nov 24 13:48:00 crc kubenswrapper[4824]: I1124 13:48:00.856936 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z58zc" podStartSLOduration=3.392757707 podStartE2EDuration="5.856919529s" podCreationTimestamp="2025-11-24 13:47:55 +0000 UTC" firstStartedPulling="2025-11-24 13:47:57.796003208 +0000 UTC m=+2019.435542518" lastFinishedPulling="2025-11-24 13:48:00.26016502 +0000 UTC m=+2021.899704340" observedRunningTime="2025-11-24 13:48:00.8506289 +0000 UTC m=+2022.490168210" watchObservedRunningTime="2025-11-24 13:48:00.856919529 +0000 UTC m=+2022.496458839" Nov 24 13:48:05 crc kubenswrapper[4824]: I1124 13:48:05.868620 4824 generic.go:334] "Generic (PLEG): container finished" podID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerID="16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077" exitCode=0 Nov 24 13:48:05 crc kubenswrapper[4824]: I1124 13:48:05.869157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdgcn" event={"ID":"974d2a77-e0fd-4a6d-8558-b8c4423777f6","Type":"ContainerDied","Data":"16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077"} Nov 24 13:48:06 crc kubenswrapper[4824]: I1124 13:48:06.471418 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:48:06 crc kubenswrapper[4824]: I1124 13:48:06.471497 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:48:06 crc kubenswrapper[4824]: I1124 13:48:06.550912 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:48:06 crc kubenswrapper[4824]: I1124 13:48:06.881396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdgcn" event={"ID":"974d2a77-e0fd-4a6d-8558-b8c4423777f6","Type":"ContainerStarted","Data":"3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba"} Nov 24 13:48:06 crc kubenswrapper[4824]: I1124 13:48:06.956925 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:48:06 crc kubenswrapper[4824]: I1124 13:48:06.980004 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bdgcn" podStartSLOduration=2.523328439 podStartE2EDuration="9.979989455s" podCreationTimestamp="2025-11-24 13:47:57 +0000 UTC" firstStartedPulling="2025-11-24 13:47:58.811013821 +0000 UTC m=+2020.450553131" lastFinishedPulling="2025-11-24 13:48:06.267674837 +0000 UTC m=+2027.907214147" observedRunningTime="2025-11-24 13:48:06.909955741 +0000 UTC m=+2028.549495051" watchObservedRunningTime="2025-11-24 13:48:06.979989455 +0000 UTC m=+2028.619528765" Nov 24 13:48:07 crc kubenswrapper[4824]: I1124 13:48:07.592258 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:48:07 crc kubenswrapper[4824]: I1124 13:48:07.592339 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:48:08 crc kubenswrapper[4824]: I1124 13:48:08.110023 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z58zc"] Nov 24 13:48:08 crc kubenswrapper[4824]: I1124 13:48:08.655475 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bdgcn" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerName="registry-server" probeResult="failure" output=< Nov 24 13:48:08 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 13:48:08 crc kubenswrapper[4824]: > Nov 24 13:48:08 crc kubenswrapper[4824]: I1124 13:48:08.897657 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z58zc" podUID="d97159cf-d145-4107-8451-5ee2a8255079" containerName="registry-server" containerID="cri-o://68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3" gracePeriod=2 Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.434449 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.616606 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-catalog-content\") pod \"d97159cf-d145-4107-8451-5ee2a8255079\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.616711 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-utilities\") pod \"d97159cf-d145-4107-8451-5ee2a8255079\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.616846 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b46ls\" (UniqueName: \"kubernetes.io/projected/d97159cf-d145-4107-8451-5ee2a8255079-kube-api-access-b46ls\") pod \"d97159cf-d145-4107-8451-5ee2a8255079\" (UID: \"d97159cf-d145-4107-8451-5ee2a8255079\") " Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.617824 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-utilities" (OuterVolumeSpecName: "utilities") pod "d97159cf-d145-4107-8451-5ee2a8255079" (UID: "d97159cf-d145-4107-8451-5ee2a8255079"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.623237 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d97159cf-d145-4107-8451-5ee2a8255079-kube-api-access-b46ls" (OuterVolumeSpecName: "kube-api-access-b46ls") pod "d97159cf-d145-4107-8451-5ee2a8255079" (UID: "d97159cf-d145-4107-8451-5ee2a8255079"). InnerVolumeSpecName "kube-api-access-b46ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.632053 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d97159cf-d145-4107-8451-5ee2a8255079" (UID: "d97159cf-d145-4107-8451-5ee2a8255079"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.720627 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b46ls\" (UniqueName: \"kubernetes.io/projected/d97159cf-d145-4107-8451-5ee2a8255079-kube-api-access-b46ls\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.720727 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.720793 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d97159cf-d145-4107-8451-5ee2a8255079-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.911834 4824 generic.go:334] "Generic (PLEG): container finished" podID="d97159cf-d145-4107-8451-5ee2a8255079" containerID="68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3" exitCode=0 Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.912288 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z58zc" Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.916908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z58zc" event={"ID":"d97159cf-d145-4107-8451-5ee2a8255079","Type":"ContainerDied","Data":"68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3"} Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.917087 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z58zc" event={"ID":"d97159cf-d145-4107-8451-5ee2a8255079","Type":"ContainerDied","Data":"20a26a86052dee32bae1745ce62d6ff566a57e5552847f654defe004cd32488a"} Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.917221 4824 scope.go:117] "RemoveContainer" containerID="68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3" Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.964145 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z58zc"] Nov 24 13:48:09 crc kubenswrapper[4824]: I1124 13:48:09.994789 4824 scope.go:117] "RemoveContainer" containerID="333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639" Nov 24 13:48:10 crc kubenswrapper[4824]: I1124 13:48:10.001795 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z58zc"] Nov 24 13:48:10 crc kubenswrapper[4824]: I1124 13:48:10.047269 4824 scope.go:117] "RemoveContainer" containerID="a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941" Nov 24 13:48:10 crc kubenswrapper[4824]: I1124 13:48:10.105022 4824 scope.go:117] "RemoveContainer" containerID="68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3" Nov 24 13:48:10 crc kubenswrapper[4824]: E1124 13:48:10.108967 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3\": container with ID starting with 68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3 not found: ID does not exist" containerID="68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3" Nov 24 13:48:10 crc kubenswrapper[4824]: I1124 13:48:10.109025 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3"} err="failed to get container status \"68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3\": rpc error: code = NotFound desc = could not find container \"68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3\": container with ID starting with 68a7037361af00714508945ebaf86d262a31d4ee35c122c4a6cd0a27106ee9b3 not found: ID does not exist" Nov 24 13:48:10 crc kubenswrapper[4824]: I1124 13:48:10.109047 4824 scope.go:117] "RemoveContainer" containerID="333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639" Nov 24 13:48:10 crc kubenswrapper[4824]: E1124 13:48:10.109443 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639\": container with ID starting with 333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639 not found: ID does not exist" containerID="333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639" Nov 24 13:48:10 crc kubenswrapper[4824]: I1124 13:48:10.109485 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639"} err="failed to get container status \"333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639\": rpc error: code = NotFound desc = could not find container \"333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639\": container with ID starting with 333f353aec7cc0ee38c9301f408c200f542e046158d6a6032b8370da0876e639 not found: ID does not exist" Nov 24 13:48:10 crc kubenswrapper[4824]: I1124 13:48:10.109513 4824 scope.go:117] "RemoveContainer" containerID="a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941" Nov 24 13:48:10 crc kubenswrapper[4824]: E1124 13:48:10.110333 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941\": container with ID starting with a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941 not found: ID does not exist" containerID="a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941" Nov 24 13:48:10 crc kubenswrapper[4824]: I1124 13:48:10.110392 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941"} err="failed to get container status \"a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941\": rpc error: code = NotFound desc = could not find container \"a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941\": container with ID starting with a638d6169eed88959e372dd6aa42fd1ab9a113bee2ab34af7a48eb9452ff2941 not found: ID does not exist" Nov 24 13:48:11 crc kubenswrapper[4824]: I1124 13:48:11.023189 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d97159cf-d145-4107-8451-5ee2a8255079" path="/var/lib/kubelet/pods/d97159cf-d145-4107-8451-5ee2a8255079/volumes" Nov 24 13:48:17 crc kubenswrapper[4824]: I1124 13:48:17.641170 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:48:17 crc kubenswrapper[4824]: I1124 13:48:17.700181 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:48:17 crc kubenswrapper[4824]: I1124 13:48:17.878977 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdgcn"] Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.045802 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bdgcn" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerName="registry-server" containerID="cri-o://3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba" gracePeriod=2 Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.553294 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.739503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-utilities\") pod \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.739868 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-catalog-content\") pod \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.740071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg7b9\" (UniqueName: \"kubernetes.io/projected/974d2a77-e0fd-4a6d-8558-b8c4423777f6-kube-api-access-mg7b9\") pod \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\" (UID: \"974d2a77-e0fd-4a6d-8558-b8c4423777f6\") " Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.740273 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-utilities" (OuterVolumeSpecName: "utilities") pod "974d2a77-e0fd-4a6d-8558-b8c4423777f6" (UID: "974d2a77-e0fd-4a6d-8558-b8c4423777f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.741153 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.750833 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/974d2a77-e0fd-4a6d-8558-b8c4423777f6-kube-api-access-mg7b9" (OuterVolumeSpecName: "kube-api-access-mg7b9") pod "974d2a77-e0fd-4a6d-8558-b8c4423777f6" (UID: "974d2a77-e0fd-4a6d-8558-b8c4423777f6"). InnerVolumeSpecName "kube-api-access-mg7b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.830978 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "974d2a77-e0fd-4a6d-8558-b8c4423777f6" (UID: "974d2a77-e0fd-4a6d-8558-b8c4423777f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.843316 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg7b9\" (UniqueName: \"kubernetes.io/projected/974d2a77-e0fd-4a6d-8558-b8c4423777f6-kube-api-access-mg7b9\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:19 crc kubenswrapper[4824]: I1124 13:48:19.843351 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974d2a77-e0fd-4a6d-8558-b8c4423777f6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.059212 4824 generic.go:334] "Generic (PLEG): container finished" podID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerID="3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba" exitCode=0 Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.059261 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdgcn" event={"ID":"974d2a77-e0fd-4a6d-8558-b8c4423777f6","Type":"ContainerDied","Data":"3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba"} Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.059290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdgcn" event={"ID":"974d2a77-e0fd-4a6d-8558-b8c4423777f6","Type":"ContainerDied","Data":"37b3b3a98f0a670a4a5919186963a92c724ff24b217acfd43a9ff8dfa39b706e"} Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.059305 4824 scope.go:117] "RemoveContainer" containerID="3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.059428 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdgcn" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.101201 4824 scope.go:117] "RemoveContainer" containerID="16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.102147 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdgcn"] Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.107387 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bdgcn"] Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.126293 4824 scope.go:117] "RemoveContainer" containerID="6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.167651 4824 scope.go:117] "RemoveContainer" containerID="3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba" Nov 24 13:48:20 crc kubenswrapper[4824]: E1124 13:48:20.168377 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba\": container with ID starting with 3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba not found: ID does not exist" containerID="3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.168406 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba"} err="failed to get container status \"3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba\": rpc error: code = NotFound desc = could not find container \"3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba\": container with ID starting with 3cb964d90ade84762bbc51dfa8d1291acf08bd95acb5af9cbc270fbf2ef0bdba not found: ID does not exist" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.168447 4824 scope.go:117] "RemoveContainer" containerID="16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077" Nov 24 13:48:20 crc kubenswrapper[4824]: E1124 13:48:20.168722 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077\": container with ID starting with 16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077 not found: ID does not exist" containerID="16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.168763 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077"} err="failed to get container status \"16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077\": rpc error: code = NotFound desc = could not find container \"16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077\": container with ID starting with 16a136669a7310947315a647b7dca70036bbfe7b5f31e3ae6dc570f4d37d2077 not found: ID does not exist" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.168778 4824 scope.go:117] "RemoveContainer" containerID="6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae" Nov 24 13:48:20 crc kubenswrapper[4824]: E1124 13:48:20.169002 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae\": container with ID starting with 6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae not found: ID does not exist" containerID="6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae" Nov 24 13:48:20 crc kubenswrapper[4824]: I1124 13:48:20.169037 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae"} err="failed to get container status \"6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae\": rpc error: code = NotFound desc = could not find container \"6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae\": container with ID starting with 6492c7e63da20f2f6970d01c824f5665f235db81c234793b25a70fc2ccfb4dae not found: ID does not exist" Nov 24 13:48:21 crc kubenswrapper[4824]: I1124 13:48:21.020028 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" path="/var/lib/kubelet/pods/974d2a77-e0fd-4a6d-8558-b8c4423777f6/volumes" Nov 24 13:48:24 crc kubenswrapper[4824]: I1124 13:48:24.109987 4824 generic.go:334] "Generic (PLEG): container finished" podID="92ed5101-5b03-4be3-bd82-32ba9915ba15" containerID="339d32cac1461e4e1a6ffdeef7fb79d571e89f8553485706170f1f510770a8ea" exitCode=0 Nov 24 13:48:24 crc kubenswrapper[4824]: I1124 13:48:24.110244 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" event={"ID":"92ed5101-5b03-4be3-bd82-32ba9915ba15","Type":"ContainerDied","Data":"339d32cac1461e4e1a6ffdeef7fb79d571e89f8553485706170f1f510770a8ea"} Nov 24 13:48:24 crc kubenswrapper[4824]: I1124 13:48:24.233366 4824 scope.go:117] "RemoveContainer" containerID="b6ba79778b2ba939a1dca97c40ae09b5ac5f5f75367717e7a9caa7280cca9c94" Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.658751 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.751278 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-inventory\") pod \"92ed5101-5b03-4be3-bd82-32ba9915ba15\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.751574 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-ssh-key\") pod \"92ed5101-5b03-4be3-bd82-32ba9915ba15\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.751896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn474\" (UniqueName: \"kubernetes.io/projected/92ed5101-5b03-4be3-bd82-32ba9915ba15-kube-api-access-gn474\") pod \"92ed5101-5b03-4be3-bd82-32ba9915ba15\" (UID: \"92ed5101-5b03-4be3-bd82-32ba9915ba15\") " Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.775092 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92ed5101-5b03-4be3-bd82-32ba9915ba15-kube-api-access-gn474" (OuterVolumeSpecName: "kube-api-access-gn474") pod "92ed5101-5b03-4be3-bd82-32ba9915ba15" (UID: "92ed5101-5b03-4be3-bd82-32ba9915ba15"). InnerVolumeSpecName "kube-api-access-gn474". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.836979 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "92ed5101-5b03-4be3-bd82-32ba9915ba15" (UID: "92ed5101-5b03-4be3-bd82-32ba9915ba15"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.854018 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn474\" (UniqueName: \"kubernetes.io/projected/92ed5101-5b03-4be3-bd82-32ba9915ba15-kube-api-access-gn474\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.854050 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.856677 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-inventory" (OuterVolumeSpecName: "inventory") pod "92ed5101-5b03-4be3-bd82-32ba9915ba15" (UID: "92ed5101-5b03-4be3-bd82-32ba9915ba15"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:48:25 crc kubenswrapper[4824]: I1124 13:48:25.955326 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92ed5101-5b03-4be3-bd82-32ba9915ba15-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.129672 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" event={"ID":"92ed5101-5b03-4be3-bd82-32ba9915ba15","Type":"ContainerDied","Data":"f2fba5d92a271e0e96b88cc10b1ac35fe5950f8a3c9092d105de105847a3b0b8"} Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.129716 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2fba5d92a271e0e96b88cc10b1ac35fe5950f8a3c9092d105de105847a3b0b8" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.130270 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.233924 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vtnl7"] Nov 24 13:48:26 crc kubenswrapper[4824]: E1124 13:48:26.234537 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97159cf-d145-4107-8451-5ee2a8255079" containerName="extract-utilities" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.234614 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97159cf-d145-4107-8451-5ee2a8255079" containerName="extract-utilities" Nov 24 13:48:26 crc kubenswrapper[4824]: E1124 13:48:26.234679 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97159cf-d145-4107-8451-5ee2a8255079" containerName="extract-content" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.234737 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97159cf-d145-4107-8451-5ee2a8255079" containerName="extract-content" Nov 24 13:48:26 crc kubenswrapper[4824]: E1124 13:48:26.234794 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerName="extract-utilities" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.234862 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerName="extract-utilities" Nov 24 13:48:26 crc kubenswrapper[4824]: E1124 13:48:26.234920 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ed5101-5b03-4be3-bd82-32ba9915ba15" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.234980 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ed5101-5b03-4be3-bd82-32ba9915ba15" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:48:26 crc kubenswrapper[4824]: E1124 13:48:26.235041 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerName="registry-server" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.235094 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerName="registry-server" Nov 24 13:48:26 crc kubenswrapper[4824]: E1124 13:48:26.235159 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerName="extract-content" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.235231 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerName="extract-content" Nov 24 13:48:26 crc kubenswrapper[4824]: E1124 13:48:26.235291 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97159cf-d145-4107-8451-5ee2a8255079" containerName="registry-server" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.235340 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97159cf-d145-4107-8451-5ee2a8255079" containerName="registry-server" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.235628 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="974d2a77-e0fd-4a6d-8558-b8c4423777f6" containerName="registry-server" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.235723 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d97159cf-d145-4107-8451-5ee2a8255079" containerName="registry-server" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.235784 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="92ed5101-5b03-4be3-bd82-32ba9915ba15" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.236480 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.238670 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.239167 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.239421 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.242057 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.244887 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vtnl7"] Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.362474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vtnl7\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.362885 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k59vn\" (UniqueName: \"kubernetes.io/projected/62b46948-d9cc-4a68-8d94-70a7538bd61d-kube-api-access-k59vn\") pod \"ssh-known-hosts-edpm-deployment-vtnl7\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.362974 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vtnl7\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.464889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vtnl7\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.464958 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k59vn\" (UniqueName: \"kubernetes.io/projected/62b46948-d9cc-4a68-8d94-70a7538bd61d-kube-api-access-k59vn\") pod \"ssh-known-hosts-edpm-deployment-vtnl7\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.465023 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vtnl7\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.470292 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vtnl7\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.470758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vtnl7\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.482575 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k59vn\" (UniqueName: \"kubernetes.io/projected/62b46948-d9cc-4a68-8d94-70a7538bd61d-kube-api-access-k59vn\") pod \"ssh-known-hosts-edpm-deployment-vtnl7\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:26 crc kubenswrapper[4824]: I1124 13:48:26.551869 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:27 crc kubenswrapper[4824]: I1124 13:48:27.149526 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vtnl7"] Nov 24 13:48:27 crc kubenswrapper[4824]: W1124 13:48:27.155138 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62b46948_d9cc_4a68_8d94_70a7538bd61d.slice/crio-006daf1141168736f079cd4eb3e74c464dde529b7a3e90c14b5d0f05427d5f79 WatchSource:0}: Error finding container 006daf1141168736f079cd4eb3e74c464dde529b7a3e90c14b5d0f05427d5f79: Status 404 returned error can't find the container with id 006daf1141168736f079cd4eb3e74c464dde529b7a3e90c14b5d0f05427d5f79 Nov 24 13:48:28 crc kubenswrapper[4824]: I1124 13:48:28.162369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" event={"ID":"62b46948-d9cc-4a68-8d94-70a7538bd61d","Type":"ContainerStarted","Data":"6a222ae1d7849a05d2f1e549ca03d04ecdc5771a06ce601e9a3602209f9a71f7"} Nov 24 13:48:28 crc kubenswrapper[4824]: I1124 13:48:28.163009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" event={"ID":"62b46948-d9cc-4a68-8d94-70a7538bd61d","Type":"ContainerStarted","Data":"006daf1141168736f079cd4eb3e74c464dde529b7a3e90c14b5d0f05427d5f79"} Nov 24 13:48:28 crc kubenswrapper[4824]: I1124 13:48:28.189452 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" podStartSLOduration=2.018601858 podStartE2EDuration="2.189432726s" podCreationTimestamp="2025-11-24 13:48:26 +0000 UTC" firstStartedPulling="2025-11-24 13:48:27.159885649 +0000 UTC m=+2048.799424989" lastFinishedPulling="2025-11-24 13:48:27.330716537 +0000 UTC m=+2048.970255857" observedRunningTime="2025-11-24 13:48:28.182269344 +0000 UTC m=+2049.821808654" watchObservedRunningTime="2025-11-24 13:48:28.189432726 +0000 UTC m=+2049.828972026" Nov 24 13:48:35 crc kubenswrapper[4824]: I1124 13:48:35.226136 4824 generic.go:334] "Generic (PLEG): container finished" podID="62b46948-d9cc-4a68-8d94-70a7538bd61d" containerID="6a222ae1d7849a05d2f1e549ca03d04ecdc5771a06ce601e9a3602209f9a71f7" exitCode=0 Nov 24 13:48:35 crc kubenswrapper[4824]: I1124 13:48:35.226200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" event={"ID":"62b46948-d9cc-4a68-8d94-70a7538bd61d","Type":"ContainerDied","Data":"6a222ae1d7849a05d2f1e549ca03d04ecdc5771a06ce601e9a3602209f9a71f7"} Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.746647 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.778582 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-ssh-key-openstack-edpm-ipam\") pod \"62b46948-d9cc-4a68-8d94-70a7538bd61d\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.778669 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-inventory-0\") pod \"62b46948-d9cc-4a68-8d94-70a7538bd61d\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.778852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k59vn\" (UniqueName: \"kubernetes.io/projected/62b46948-d9cc-4a68-8d94-70a7538bd61d-kube-api-access-k59vn\") pod \"62b46948-d9cc-4a68-8d94-70a7538bd61d\" (UID: \"62b46948-d9cc-4a68-8d94-70a7538bd61d\") " Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.783926 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b46948-d9cc-4a68-8d94-70a7538bd61d-kube-api-access-k59vn" (OuterVolumeSpecName: "kube-api-access-k59vn") pod "62b46948-d9cc-4a68-8d94-70a7538bd61d" (UID: "62b46948-d9cc-4a68-8d94-70a7538bd61d"). InnerVolumeSpecName "kube-api-access-k59vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.807063 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "62b46948-d9cc-4a68-8d94-70a7538bd61d" (UID: "62b46948-d9cc-4a68-8d94-70a7538bd61d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.808030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "62b46948-d9cc-4a68-8d94-70a7538bd61d" (UID: "62b46948-d9cc-4a68-8d94-70a7538bd61d"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.881756 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.881781 4824 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/62b46948-d9cc-4a68-8d94-70a7538bd61d-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:36 crc kubenswrapper[4824]: I1124 13:48:36.881790 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k59vn\" (UniqueName: \"kubernetes.io/projected/62b46948-d9cc-4a68-8d94-70a7538bd61d-kube-api-access-k59vn\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.280785 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" event={"ID":"62b46948-d9cc-4a68-8d94-70a7538bd61d","Type":"ContainerDied","Data":"006daf1141168736f079cd4eb3e74c464dde529b7a3e90c14b5d0f05427d5f79"} Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.280846 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="006daf1141168736f079cd4eb3e74c464dde529b7a3e90c14b5d0f05427d5f79" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.280947 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vtnl7" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.346388 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l"] Nov 24 13:48:37 crc kubenswrapper[4824]: E1124 13:48:37.346820 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b46948-d9cc-4a68-8d94-70a7538bd61d" containerName="ssh-known-hosts-edpm-deployment" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.346857 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b46948-d9cc-4a68-8d94-70a7538bd61d" containerName="ssh-known-hosts-edpm-deployment" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.347043 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b46948-d9cc-4a68-8d94-70a7538bd61d" containerName="ssh-known-hosts-edpm-deployment" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.347667 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.350323 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.352846 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.352844 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.352952 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.372351 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l"] Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.389536 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kfht\" (UniqueName: \"kubernetes.io/projected/9ed4df3b-472c-4815-a5ef-66125b2dde81-kube-api-access-8kfht\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-97k6l\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.389669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-97k6l\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.389788 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-97k6l\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.490936 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-97k6l\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.491014 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kfht\" (UniqueName: \"kubernetes.io/projected/9ed4df3b-472c-4815-a5ef-66125b2dde81-kube-api-access-8kfht\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-97k6l\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.491111 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-97k6l\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.496662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-97k6l\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.499480 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-97k6l\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.512394 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kfht\" (UniqueName: \"kubernetes.io/projected/9ed4df3b-472c-4815-a5ef-66125b2dde81-kube-api-access-8kfht\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-97k6l\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:37 crc kubenswrapper[4824]: I1124 13:48:37.662520 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:38 crc kubenswrapper[4824]: I1124 13:48:38.251611 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l"] Nov 24 13:48:38 crc kubenswrapper[4824]: I1124 13:48:38.296047 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" event={"ID":"9ed4df3b-472c-4815-a5ef-66125b2dde81","Type":"ContainerStarted","Data":"7be3cd1a76da752b1134e6878aea4d26bb24e5ceb8152bf87523386dfa5fd095"} Nov 24 13:48:39 crc kubenswrapper[4824]: I1124 13:48:39.309141 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" event={"ID":"9ed4df3b-472c-4815-a5ef-66125b2dde81","Type":"ContainerStarted","Data":"941a100ccc533a5c590bdea4adb094051f7ede7a6e9e3ab25617bb6758673fc3"} Nov 24 13:48:39 crc kubenswrapper[4824]: I1124 13:48:39.336332 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" podStartSLOduration=2.18062376 podStartE2EDuration="2.336307034s" podCreationTimestamp="2025-11-24 13:48:37 +0000 UTC" firstStartedPulling="2025-11-24 13:48:38.264590569 +0000 UTC m=+2059.904129869" lastFinishedPulling="2025-11-24 13:48:38.420273803 +0000 UTC m=+2060.059813143" observedRunningTime="2025-11-24 13:48:39.32785265 +0000 UTC m=+2060.967392030" watchObservedRunningTime="2025-11-24 13:48:39.336307034 +0000 UTC m=+2060.975846344" Nov 24 13:48:47 crc kubenswrapper[4824]: I1124 13:48:47.393793 4824 generic.go:334] "Generic (PLEG): container finished" podID="9ed4df3b-472c-4815-a5ef-66125b2dde81" containerID="941a100ccc533a5c590bdea4adb094051f7ede7a6e9e3ab25617bb6758673fc3" exitCode=0 Nov 24 13:48:47 crc kubenswrapper[4824]: I1124 13:48:47.393956 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" event={"ID":"9ed4df3b-472c-4815-a5ef-66125b2dde81","Type":"ContainerDied","Data":"941a100ccc533a5c590bdea4adb094051f7ede7a6e9e3ab25617bb6758673fc3"} Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.131189 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.295652 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kfht\" (UniqueName: \"kubernetes.io/projected/9ed4df3b-472c-4815-a5ef-66125b2dde81-kube-api-access-8kfht\") pod \"9ed4df3b-472c-4815-a5ef-66125b2dde81\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.295782 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-inventory\") pod \"9ed4df3b-472c-4815-a5ef-66125b2dde81\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.295863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-ssh-key\") pod \"9ed4df3b-472c-4815-a5ef-66125b2dde81\" (UID: \"9ed4df3b-472c-4815-a5ef-66125b2dde81\") " Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.306102 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed4df3b-472c-4815-a5ef-66125b2dde81-kube-api-access-8kfht" (OuterVolumeSpecName: "kube-api-access-8kfht") pod "9ed4df3b-472c-4815-a5ef-66125b2dde81" (UID: "9ed4df3b-472c-4815-a5ef-66125b2dde81"). InnerVolumeSpecName "kube-api-access-8kfht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.329008 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-inventory" (OuterVolumeSpecName: "inventory") pod "9ed4df3b-472c-4815-a5ef-66125b2dde81" (UID: "9ed4df3b-472c-4815-a5ef-66125b2dde81"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.332586 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9ed4df3b-472c-4815-a5ef-66125b2dde81" (UID: "9ed4df3b-472c-4815-a5ef-66125b2dde81"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.398472 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kfht\" (UniqueName: \"kubernetes.io/projected/9ed4df3b-472c-4815-a5ef-66125b2dde81-kube-api-access-8kfht\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.398519 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.398539 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ed4df3b-472c-4815-a5ef-66125b2dde81-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.726793 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" event={"ID":"9ed4df3b-472c-4815-a5ef-66125b2dde81","Type":"ContainerDied","Data":"7be3cd1a76da752b1134e6878aea4d26bb24e5ceb8152bf87523386dfa5fd095"} Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.727055 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7be3cd1a76da752b1134e6878aea4d26bb24e5ceb8152bf87523386dfa5fd095" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.726881 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-97k6l" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.803559 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h"] Nov 24 13:48:49 crc kubenswrapper[4824]: E1124 13:48:49.804312 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed4df3b-472c-4815-a5ef-66125b2dde81" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.804404 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed4df3b-472c-4815-a5ef-66125b2dde81" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.804783 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed4df3b-472c-4815-a5ef-66125b2dde81" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.805707 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.809567 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.809925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz5ps\" (UniqueName: \"kubernetes.io/projected/20a6cb74-657c-42ba-a789-411e312d605b-kube-api-access-zz5ps\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.810072 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.813507 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.813939 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.814086 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.814241 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.821761 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h"] Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.912401 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz5ps\" (UniqueName: \"kubernetes.io/projected/20a6cb74-657c-42ba-a789-411e312d605b-kube-api-access-zz5ps\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.912495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.912580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.919371 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.922523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:49 crc kubenswrapper[4824]: I1124 13:48:49.939553 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz5ps\" (UniqueName: \"kubernetes.io/projected/20a6cb74-657c-42ba-a789-411e312d605b-kube-api-access-zz5ps\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:50 crc kubenswrapper[4824]: I1124 13:48:50.142294 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:48:50 crc kubenswrapper[4824]: I1124 13:48:50.781095 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h"] Nov 24 13:48:51 crc kubenswrapper[4824]: I1124 13:48:51.744770 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" event={"ID":"20a6cb74-657c-42ba-a789-411e312d605b","Type":"ContainerStarted","Data":"17c907453862a70351df0bfd094d3d27242ce47aff29dc2649034e89fbb0896c"} Nov 24 13:48:51 crc kubenswrapper[4824]: I1124 13:48:51.745219 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" event={"ID":"20a6cb74-657c-42ba-a789-411e312d605b","Type":"ContainerStarted","Data":"cb6c8bc7a885471cbdc74ff4059db1688fcee0811581428c303687bfcb31d1fe"} Nov 24 13:48:51 crc kubenswrapper[4824]: I1124 13:48:51.769324 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" podStartSLOduration=2.59147022 podStartE2EDuration="2.769301725s" podCreationTimestamp="2025-11-24 13:48:49 +0000 UTC" firstStartedPulling="2025-11-24 13:48:50.795170389 +0000 UTC m=+2072.434709699" lastFinishedPulling="2025-11-24 13:48:50.973001894 +0000 UTC m=+2072.612541204" observedRunningTime="2025-11-24 13:48:51.765018017 +0000 UTC m=+2073.404557347" watchObservedRunningTime="2025-11-24 13:48:51.769301725 +0000 UTC m=+2073.408841085" Nov 24 13:49:01 crc kubenswrapper[4824]: I1124 13:49:01.848390 4824 generic.go:334] "Generic (PLEG): container finished" podID="20a6cb74-657c-42ba-a789-411e312d605b" containerID="17c907453862a70351df0bfd094d3d27242ce47aff29dc2649034e89fbb0896c" exitCode=0 Nov 24 13:49:01 crc kubenswrapper[4824]: I1124 13:49:01.848483 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" event={"ID":"20a6cb74-657c-42ba-a789-411e312d605b","Type":"ContainerDied","Data":"17c907453862a70351df0bfd094d3d27242ce47aff29dc2649034e89fbb0896c"} Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.359171 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.504061 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-inventory\") pod \"20a6cb74-657c-42ba-a789-411e312d605b\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.504300 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-ssh-key\") pod \"20a6cb74-657c-42ba-a789-411e312d605b\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.504347 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz5ps\" (UniqueName: \"kubernetes.io/projected/20a6cb74-657c-42ba-a789-411e312d605b-kube-api-access-zz5ps\") pod \"20a6cb74-657c-42ba-a789-411e312d605b\" (UID: \"20a6cb74-657c-42ba-a789-411e312d605b\") " Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.510950 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a6cb74-657c-42ba-a789-411e312d605b-kube-api-access-zz5ps" (OuterVolumeSpecName: "kube-api-access-zz5ps") pod "20a6cb74-657c-42ba-a789-411e312d605b" (UID: "20a6cb74-657c-42ba-a789-411e312d605b"). InnerVolumeSpecName "kube-api-access-zz5ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.535069 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "20a6cb74-657c-42ba-a789-411e312d605b" (UID: "20a6cb74-657c-42ba-a789-411e312d605b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.535526 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-inventory" (OuterVolumeSpecName: "inventory") pod "20a6cb74-657c-42ba-a789-411e312d605b" (UID: "20a6cb74-657c-42ba-a789-411e312d605b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.606775 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.606819 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz5ps\" (UniqueName: \"kubernetes.io/projected/20a6cb74-657c-42ba-a789-411e312d605b-kube-api-access-zz5ps\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.606832 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20a6cb74-657c-42ba-a789-411e312d605b-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.865791 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" event={"ID":"20a6cb74-657c-42ba-a789-411e312d605b","Type":"ContainerDied","Data":"cb6c8bc7a885471cbdc74ff4059db1688fcee0811581428c303687bfcb31d1fe"} Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.865849 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb6c8bc7a885471cbdc74ff4059db1688fcee0811581428c303687bfcb31d1fe" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.865930 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.975105 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv"] Nov 24 13:49:03 crc kubenswrapper[4824]: E1124 13:49:03.975470 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a6cb74-657c-42ba-a789-411e312d605b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.975487 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a6cb74-657c-42ba-a789-411e312d605b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.975690 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a6cb74-657c-42ba-a789-411e312d605b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.982613 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.989158 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.989325 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.989438 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.989478 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.989478 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.990892 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.992386 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.992588 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 24 13:49:03 crc kubenswrapper[4824]: I1124 13:49:03.997396 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv"] Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.122231 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.122306 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.122375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.122420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.122463 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.122552 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.122743 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.122927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.123016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.123061 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.123097 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-kube-api-access-m8gfw\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.123223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.123281 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.123332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.225923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226022 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226158 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-kube-api-access-m8gfw\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226468 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226506 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226590 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.226715 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.230278 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.230737 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.232279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.232952 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.233056 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.233488 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.234797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.235424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.235621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.236771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.237484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.247072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.251234 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-kube-api-access-m8gfw\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.251952 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.301598 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:04 crc kubenswrapper[4824]: W1124 13:49:04.879261 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27890aad_2f36_400e_8535_afecafb22a37.slice/crio-08d0cdbdc396f8c887e3f4fa5e282087c8d1dc4381c13b7a960a549483d752d9 WatchSource:0}: Error finding container 08d0cdbdc396f8c887e3f4fa5e282087c8d1dc4381c13b7a960a549483d752d9: Status 404 returned error can't find the container with id 08d0cdbdc396f8c887e3f4fa5e282087c8d1dc4381c13b7a960a549483d752d9 Nov 24 13:49:04 crc kubenswrapper[4824]: I1124 13:49:04.880433 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv"] Nov 24 13:49:05 crc kubenswrapper[4824]: I1124 13:49:05.883438 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" event={"ID":"27890aad-2f36-400e-8535-afecafb22a37","Type":"ContainerStarted","Data":"d17dc16e0c1a190ce011f0c1a9ffb7ddbe3d1d1e6a20f9afab79ce96da5f97ec"} Nov 24 13:49:05 crc kubenswrapper[4824]: I1124 13:49:05.883778 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" event={"ID":"27890aad-2f36-400e-8535-afecafb22a37","Type":"ContainerStarted","Data":"08d0cdbdc396f8c887e3f4fa5e282087c8d1dc4381c13b7a960a549483d752d9"} Nov 24 13:49:05 crc kubenswrapper[4824]: I1124 13:49:05.915185 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" podStartSLOduration=2.7357045700000002 podStartE2EDuration="2.915161745s" podCreationTimestamp="2025-11-24 13:49:03 +0000 UTC" firstStartedPulling="2025-11-24 13:49:04.889126781 +0000 UTC m=+2086.528666111" lastFinishedPulling="2025-11-24 13:49:05.068583976 +0000 UTC m=+2086.708123286" observedRunningTime="2025-11-24 13:49:05.90467851 +0000 UTC m=+2087.544217840" watchObservedRunningTime="2025-11-24 13:49:05.915161745 +0000 UTC m=+2087.554701075" Nov 24 13:49:31 crc kubenswrapper[4824]: I1124 13:49:31.770529 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vk6jr"] Nov 24 13:49:31 crc kubenswrapper[4824]: I1124 13:49:31.775898 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:31 crc kubenswrapper[4824]: I1124 13:49:31.805095 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vk6jr"] Nov 24 13:49:31 crc kubenswrapper[4824]: I1124 13:49:31.966966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-catalog-content\") pod \"certified-operators-vk6jr\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:31 crc kubenswrapper[4824]: I1124 13:49:31.967958 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd8j7\" (UniqueName: \"kubernetes.io/projected/1d130b75-322c-4791-8d11-d99107f41053-kube-api-access-hd8j7\") pod \"certified-operators-vk6jr\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:31 crc kubenswrapper[4824]: I1124 13:49:31.968228 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-utilities\") pod \"certified-operators-vk6jr\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:32 crc kubenswrapper[4824]: I1124 13:49:32.069920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-catalog-content\") pod \"certified-operators-vk6jr\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:32 crc kubenswrapper[4824]: I1124 13:49:32.070250 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd8j7\" (UniqueName: \"kubernetes.io/projected/1d130b75-322c-4791-8d11-d99107f41053-kube-api-access-hd8j7\") pod \"certified-operators-vk6jr\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:32 crc kubenswrapper[4824]: I1124 13:49:32.070390 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-catalog-content\") pod \"certified-operators-vk6jr\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:32 crc kubenswrapper[4824]: I1124 13:49:32.070954 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-utilities\") pod \"certified-operators-vk6jr\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:32 crc kubenswrapper[4824]: I1124 13:49:32.071299 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-utilities\") pod \"certified-operators-vk6jr\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:32 crc kubenswrapper[4824]: I1124 13:49:32.110886 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd8j7\" (UniqueName: \"kubernetes.io/projected/1d130b75-322c-4791-8d11-d99107f41053-kube-api-access-hd8j7\") pod \"certified-operators-vk6jr\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:32 crc kubenswrapper[4824]: I1124 13:49:32.401737 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:32 crc kubenswrapper[4824]: I1124 13:49:32.883528 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vk6jr"] Nov 24 13:49:33 crc kubenswrapper[4824]: I1124 13:49:33.157019 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d130b75-322c-4791-8d11-d99107f41053" containerID="579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920" exitCode=0 Nov 24 13:49:33 crc kubenswrapper[4824]: I1124 13:49:33.157414 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vk6jr" event={"ID":"1d130b75-322c-4791-8d11-d99107f41053","Type":"ContainerDied","Data":"579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920"} Nov 24 13:49:33 crc kubenswrapper[4824]: I1124 13:49:33.157458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vk6jr" event={"ID":"1d130b75-322c-4791-8d11-d99107f41053","Type":"ContainerStarted","Data":"1f138b6148cc06d4267e7bb07e7dc1e2b35d9dde88caf9faafd8eddb19cd8ce5"} Nov 24 13:49:34 crc kubenswrapper[4824]: I1124 13:49:34.168500 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vk6jr" event={"ID":"1d130b75-322c-4791-8d11-d99107f41053","Type":"ContainerStarted","Data":"d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7"} Nov 24 13:49:35 crc kubenswrapper[4824]: I1124 13:49:35.179203 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d130b75-322c-4791-8d11-d99107f41053" containerID="d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7" exitCode=0 Nov 24 13:49:35 crc kubenswrapper[4824]: I1124 13:49:35.179246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vk6jr" event={"ID":"1d130b75-322c-4791-8d11-d99107f41053","Type":"ContainerDied","Data":"d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7"} Nov 24 13:49:36 crc kubenswrapper[4824]: I1124 13:49:36.193304 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vk6jr" event={"ID":"1d130b75-322c-4791-8d11-d99107f41053","Type":"ContainerStarted","Data":"36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970"} Nov 24 13:49:36 crc kubenswrapper[4824]: I1124 13:49:36.224257 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vk6jr" podStartSLOduration=2.8073670809999998 podStartE2EDuration="5.22423246s" podCreationTimestamp="2025-11-24 13:49:31 +0000 UTC" firstStartedPulling="2025-11-24 13:49:33.15936602 +0000 UTC m=+2114.798905340" lastFinishedPulling="2025-11-24 13:49:35.576231409 +0000 UTC m=+2117.215770719" observedRunningTime="2025-11-24 13:49:36.217431659 +0000 UTC m=+2117.856970989" watchObservedRunningTime="2025-11-24 13:49:36.22423246 +0000 UTC m=+2117.863771780" Nov 24 13:49:42 crc kubenswrapper[4824]: I1124 13:49:42.402561 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:42 crc kubenswrapper[4824]: I1124 13:49:42.403178 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:42 crc kubenswrapper[4824]: I1124 13:49:42.454630 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:43 crc kubenswrapper[4824]: I1124 13:49:43.325373 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:43 crc kubenswrapper[4824]: I1124 13:49:43.404362 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vk6jr"] Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.278436 4824 generic.go:334] "Generic (PLEG): container finished" podID="27890aad-2f36-400e-8535-afecafb22a37" containerID="d17dc16e0c1a190ce011f0c1a9ffb7ddbe3d1d1e6a20f9afab79ce96da5f97ec" exitCode=0 Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.278516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" event={"ID":"27890aad-2f36-400e-8535-afecafb22a37","Type":"ContainerDied","Data":"d17dc16e0c1a190ce011f0c1a9ffb7ddbe3d1d1e6a20f9afab79ce96da5f97ec"} Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.279166 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vk6jr" podUID="1d130b75-322c-4791-8d11-d99107f41053" containerName="registry-server" containerID="cri-o://36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970" gracePeriod=2 Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.775212 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.882119 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-catalog-content\") pod \"1d130b75-322c-4791-8d11-d99107f41053\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.882276 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd8j7\" (UniqueName: \"kubernetes.io/projected/1d130b75-322c-4791-8d11-d99107f41053-kube-api-access-hd8j7\") pod \"1d130b75-322c-4791-8d11-d99107f41053\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.882368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-utilities\") pod \"1d130b75-322c-4791-8d11-d99107f41053\" (UID: \"1d130b75-322c-4791-8d11-d99107f41053\") " Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.883444 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-utilities" (OuterVolumeSpecName: "utilities") pod "1d130b75-322c-4791-8d11-d99107f41053" (UID: "1d130b75-322c-4791-8d11-d99107f41053"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.888490 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d130b75-322c-4791-8d11-d99107f41053-kube-api-access-hd8j7" (OuterVolumeSpecName: "kube-api-access-hd8j7") pod "1d130b75-322c-4791-8d11-d99107f41053" (UID: "1d130b75-322c-4791-8d11-d99107f41053"). InnerVolumeSpecName "kube-api-access-hd8j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.936466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d130b75-322c-4791-8d11-d99107f41053" (UID: "1d130b75-322c-4791-8d11-d99107f41053"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.984557 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.984592 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd8j7\" (UniqueName: \"kubernetes.io/projected/1d130b75-322c-4791-8d11-d99107f41053-kube-api-access-hd8j7\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:45 crc kubenswrapper[4824]: I1124 13:49:45.984605 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d130b75-322c-4791-8d11-d99107f41053-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.295431 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d130b75-322c-4791-8d11-d99107f41053" containerID="36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970" exitCode=0 Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.295693 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vk6jr" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.295719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vk6jr" event={"ID":"1d130b75-322c-4791-8d11-d99107f41053","Type":"ContainerDied","Data":"36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970"} Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.296933 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vk6jr" event={"ID":"1d130b75-322c-4791-8d11-d99107f41053","Type":"ContainerDied","Data":"1f138b6148cc06d4267e7bb07e7dc1e2b35d9dde88caf9faafd8eddb19cd8ce5"} Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.296983 4824 scope.go:117] "RemoveContainer" containerID="36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.363143 4824 scope.go:117] "RemoveContainer" containerID="d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.368480 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vk6jr"] Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.379653 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vk6jr"] Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.398478 4824 scope.go:117] "RemoveContainer" containerID="579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.450250 4824 scope.go:117] "RemoveContainer" containerID="36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970" Nov 24 13:49:46 crc kubenswrapper[4824]: E1124 13:49:46.450708 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970\": container with ID starting with 36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970 not found: ID does not exist" containerID="36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.450764 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970"} err="failed to get container status \"36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970\": rpc error: code = NotFound desc = could not find container \"36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970\": container with ID starting with 36e4e9560d46bf934e64758887d0663dd6dd01384a8f3ff5e8c565284a835970 not found: ID does not exist" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.450801 4824 scope.go:117] "RemoveContainer" containerID="d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7" Nov 24 13:49:46 crc kubenswrapper[4824]: E1124 13:49:46.454257 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7\": container with ID starting with d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7 not found: ID does not exist" containerID="d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.454328 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7"} err="failed to get container status \"d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7\": rpc error: code = NotFound desc = could not find container \"d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7\": container with ID starting with d9ef8c95e9b7592b448c2b8be428f7d8bb63834aca3730a46c77d7baf06637d7 not found: ID does not exist" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.454353 4824 scope.go:117] "RemoveContainer" containerID="579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920" Nov 24 13:49:46 crc kubenswrapper[4824]: E1124 13:49:46.454693 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920\": container with ID starting with 579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920 not found: ID does not exist" containerID="579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.454736 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920"} err="failed to get container status \"579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920\": rpc error: code = NotFound desc = could not find container \"579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920\": container with ID starting with 579792c76df535cf210c2b3babe9090a5a6708affddd119a35092c1e5bd6e920 not found: ID does not exist" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.733685 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-bootstrap-combined-ca-bundle\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900422 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-inventory\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900450 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-telemetry-combined-ca-bundle\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900536 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-ovn-default-certs-0\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900594 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ssh-key\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900646 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900679 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-nova-combined-ca-bundle\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900705 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-libvirt-combined-ca-bundle\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900732 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-kube-api-access-m8gfw\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900761 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900790 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ovn-combined-ca-bundle\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900839 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900861 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-repo-setup-combined-ca-bundle\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.900930 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-neutron-metadata-combined-ca-bundle\") pod \"27890aad-2f36-400e-8535-afecafb22a37\" (UID: \"27890aad-2f36-400e-8535-afecafb22a37\") " Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.911960 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.912409 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.915223 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.916850 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.916946 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.917112 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.926276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.927933 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.927996 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.928263 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.930970 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-kube-api-access-m8gfw" (OuterVolumeSpecName: "kube-api-access-m8gfw") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "kube-api-access-m8gfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.933285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.950560 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:46 crc kubenswrapper[4824]: I1124 13:49:46.962163 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-inventory" (OuterVolumeSpecName: "inventory") pod "27890aad-2f36-400e-8535-afecafb22a37" (UID: "27890aad-2f36-400e-8535-afecafb22a37"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.003604 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.003995 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-kube-api-access-m8gfw\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.004172 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.004311 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.005416 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.005609 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.005787 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.006048 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.006463 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.006635 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.006847 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.007024 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.007457 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27890aad-2f36-400e-8535-afecafb22a37-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.007656 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27890aad-2f36-400e-8535-afecafb22a37-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.025871 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d130b75-322c-4791-8d11-d99107f41053" path="/var/lib/kubelet/pods/1d130b75-322c-4791-8d11-d99107f41053/volumes" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.307408 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" event={"ID":"27890aad-2f36-400e-8535-afecafb22a37","Type":"ContainerDied","Data":"08d0cdbdc396f8c887e3f4fa5e282087c8d1dc4381c13b7a960a549483d752d9"} Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.307462 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08d0cdbdc396f8c887e3f4fa5e282087c8d1dc4381c13b7a960a549483d752d9" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.307518 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.440999 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7"] Nov 24 13:49:47 crc kubenswrapper[4824]: E1124 13:49:47.441671 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27890aad-2f36-400e-8535-afecafb22a37" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.441777 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="27890aad-2f36-400e-8535-afecafb22a37" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 13:49:47 crc kubenswrapper[4824]: E1124 13:49:47.441884 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d130b75-322c-4791-8d11-d99107f41053" containerName="extract-content" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.442020 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d130b75-322c-4791-8d11-d99107f41053" containerName="extract-content" Nov 24 13:49:47 crc kubenswrapper[4824]: E1124 13:49:47.442151 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d130b75-322c-4791-8d11-d99107f41053" containerName="extract-utilities" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.442244 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d130b75-322c-4791-8d11-d99107f41053" containerName="extract-utilities" Nov 24 13:49:47 crc kubenswrapper[4824]: E1124 13:49:47.442369 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d130b75-322c-4791-8d11-d99107f41053" containerName="registry-server" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.442469 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d130b75-322c-4791-8d11-d99107f41053" containerName="registry-server" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.442886 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="27890aad-2f36-400e-8535-afecafb22a37" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.443030 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d130b75-322c-4791-8d11-d99107f41053" containerName="registry-server" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.444003 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.447055 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.447139 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.447454 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.447504 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.447668 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.493323 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7"] Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.522510 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.522592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.522643 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz2l6\" (UniqueName: \"kubernetes.io/projected/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-kube-api-access-bz2l6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.522663 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.522689 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.782959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.783036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.783057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz2l6\" (UniqueName: \"kubernetes.io/projected/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-kube-api-access-bz2l6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.783074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.783096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.786002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.791758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.796621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.812762 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:47 crc kubenswrapper[4824]: I1124 13:49:47.822153 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz2l6\" (UniqueName: \"kubernetes.io/projected/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-kube-api-access-bz2l6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bwlv7\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:48 crc kubenswrapper[4824]: I1124 13:49:48.063228 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:49:48 crc kubenswrapper[4824]: I1124 13:49:48.644455 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7"] Nov 24 13:49:49 crc kubenswrapper[4824]: I1124 13:49:49.324396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" event={"ID":"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63","Type":"ContainerStarted","Data":"ad4d9cb3388052e00955c7c066422104bacfca25995074f4d9b7f06ec31adcbb"} Nov 24 13:49:49 crc kubenswrapper[4824]: I1124 13:49:49.324717 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" event={"ID":"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63","Type":"ContainerStarted","Data":"4108168d5ab928d687e06889e52275a6391ebd03c4ab2c9d4da9d8d5e98e7c2a"} Nov 24 13:49:49 crc kubenswrapper[4824]: I1124 13:49:49.347000 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" podStartSLOduration=2.141535818 podStartE2EDuration="2.346974929s" podCreationTimestamp="2025-11-24 13:49:47 +0000 UTC" firstStartedPulling="2025-11-24 13:49:48.653414789 +0000 UTC m=+2130.292954129" lastFinishedPulling="2025-11-24 13:49:48.85885391 +0000 UTC m=+2130.498393240" observedRunningTime="2025-11-24 13:49:49.344942598 +0000 UTC m=+2130.984481948" watchObservedRunningTime="2025-11-24 13:49:49.346974929 +0000 UTC m=+2130.986514269" Nov 24 13:50:10 crc kubenswrapper[4824]: I1124 13:50:10.787821 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:50:10 crc kubenswrapper[4824]: I1124 13:50:10.788542 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:50:40 crc kubenswrapper[4824]: I1124 13:50:40.788410 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:50:40 crc kubenswrapper[4824]: I1124 13:50:40.789060 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:50:54 crc kubenswrapper[4824]: I1124 13:50:54.964302 4824 generic.go:334] "Generic (PLEG): container finished" podID="f5eaa6e7-da63-4213-ac5d-57fac9f6cd63" containerID="ad4d9cb3388052e00955c7c066422104bacfca25995074f4d9b7f06ec31adcbb" exitCode=0 Nov 24 13:50:54 crc kubenswrapper[4824]: I1124 13:50:54.964411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" event={"ID":"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63","Type":"ContainerDied","Data":"ad4d9cb3388052e00955c7c066422104bacfca25995074f4d9b7f06ec31adcbb"} Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.394131 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.425936 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bz2l6\" (UniqueName: \"kubernetes.io/projected/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-kube-api-access-bz2l6\") pod \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.426091 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovncontroller-config-0\") pod \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.426240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ssh-key\") pod \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.426298 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-inventory\") pod \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.426395 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovn-combined-ca-bundle\") pod \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\" (UID: \"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63\") " Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.434802 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63" (UID: "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.434887 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-kube-api-access-bz2l6" (OuterVolumeSpecName: "kube-api-access-bz2l6") pod "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63" (UID: "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63"). InnerVolumeSpecName "kube-api-access-bz2l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.468924 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-inventory" (OuterVolumeSpecName: "inventory") pod "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63" (UID: "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.470058 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63" (UID: "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.471148 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63" (UID: "f5eaa6e7-da63-4213-ac5d-57fac9f6cd63"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.529076 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.529118 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.529133 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.529151 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bz2l6\" (UniqueName: \"kubernetes.io/projected/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-kube-api-access-bz2l6\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.529164 4824 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f5eaa6e7-da63-4213-ac5d-57fac9f6cd63-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.986137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" event={"ID":"f5eaa6e7-da63-4213-ac5d-57fac9f6cd63","Type":"ContainerDied","Data":"4108168d5ab928d687e06889e52275a6391ebd03c4ab2c9d4da9d8d5e98e7c2a"} Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.986183 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4108168d5ab928d687e06889e52275a6391ebd03c4ab2c9d4da9d8d5e98e7c2a" Nov 24 13:50:56 crc kubenswrapper[4824]: I1124 13:50:56.986195 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bwlv7" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.132915 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt"] Nov 24 13:50:57 crc kubenswrapper[4824]: E1124 13:50:57.134763 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5eaa6e7-da63-4213-ac5d-57fac9f6cd63" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.134800 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5eaa6e7-da63-4213-ac5d-57fac9f6cd63" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.135154 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5eaa6e7-da63-4213-ac5d-57fac9f6cd63" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.136198 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.142741 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.142938 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.143127 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.143585 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.143704 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.143885 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.146235 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt"] Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.242682 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlv7c\" (UniqueName: \"kubernetes.io/projected/18802c46-869d-4e00-848e-79df8665851d-kube-api-access-zlv7c\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.242859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.242960 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.243023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.243266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.243365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.345144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.345226 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.345285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.345382 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.345461 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.345718 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlv7c\" (UniqueName: \"kubernetes.io/projected/18802c46-869d-4e00-848e-79df8665851d-kube-api-access-zlv7c\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.349373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.349385 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.349519 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.350256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.351069 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.372267 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlv7c\" (UniqueName: \"kubernetes.io/projected/18802c46-869d-4e00-848e-79df8665851d-kube-api-access-zlv7c\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:57 crc kubenswrapper[4824]: I1124 13:50:57.466177 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:50:58 crc kubenswrapper[4824]: I1124 13:50:58.059153 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt"] Nov 24 13:50:59 crc kubenswrapper[4824]: I1124 13:50:59.032853 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" event={"ID":"18802c46-869d-4e00-848e-79df8665851d","Type":"ContainerStarted","Data":"82d11715c473edcc363a6e3392afc414dbf45b7d4125214c9869af36eb549af4"} Nov 24 13:50:59 crc kubenswrapper[4824]: I1124 13:50:59.033593 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" event={"ID":"18802c46-869d-4e00-848e-79df8665851d","Type":"ContainerStarted","Data":"234d83d9fd9d21e66bfc5f360aa4ade8822ecb034b52cceb394d39cc2428ccd2"} Nov 24 13:50:59 crc kubenswrapper[4824]: I1124 13:50:59.097576 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" podStartSLOduration=1.8822943250000002 podStartE2EDuration="2.097549634s" podCreationTimestamp="2025-11-24 13:50:57 +0000 UTC" firstStartedPulling="2025-11-24 13:50:58.071106658 +0000 UTC m=+2199.710645978" lastFinishedPulling="2025-11-24 13:50:58.286361947 +0000 UTC m=+2199.925901287" observedRunningTime="2025-11-24 13:50:59.091196683 +0000 UTC m=+2200.730736023" watchObservedRunningTime="2025-11-24 13:50:59.097549634 +0000 UTC m=+2200.737088964" Nov 24 13:51:10 crc kubenswrapper[4824]: I1124 13:51:10.787699 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:51:10 crc kubenswrapper[4824]: I1124 13:51:10.788214 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:51:10 crc kubenswrapper[4824]: I1124 13:51:10.788253 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:51:10 crc kubenswrapper[4824]: I1124 13:51:10.788990 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f7c32c7c1ab5c79ebcd1b5d0a45adaba8003211cdde83e67ef3cbc6580aa7fd2"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:51:10 crc kubenswrapper[4824]: I1124 13:51:10.789050 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://f7c32c7c1ab5c79ebcd1b5d0a45adaba8003211cdde83e67ef3cbc6580aa7fd2" gracePeriod=600 Nov 24 13:51:11 crc kubenswrapper[4824]: I1124 13:51:11.139878 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="f7c32c7c1ab5c79ebcd1b5d0a45adaba8003211cdde83e67ef3cbc6580aa7fd2" exitCode=0 Nov 24 13:51:11 crc kubenswrapper[4824]: I1124 13:51:11.140287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"f7c32c7c1ab5c79ebcd1b5d0a45adaba8003211cdde83e67ef3cbc6580aa7fd2"} Nov 24 13:51:11 crc kubenswrapper[4824]: I1124 13:51:11.140451 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b"} Nov 24 13:51:11 crc kubenswrapper[4824]: I1124 13:51:11.140476 4824 scope.go:117] "RemoveContainer" containerID="184b7e599c3658831e013de170c3d7a265bc1bee00d9e64996b12d9de6c32a80" Nov 24 13:51:48 crc kubenswrapper[4824]: I1124 13:51:48.486366 4824 generic.go:334] "Generic (PLEG): container finished" podID="18802c46-869d-4e00-848e-79df8665851d" containerID="82d11715c473edcc363a6e3392afc414dbf45b7d4125214c9869af36eb549af4" exitCode=0 Nov 24 13:51:48 crc kubenswrapper[4824]: I1124 13:51:48.486461 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" event={"ID":"18802c46-869d-4e00-848e-79df8665851d","Type":"ContainerDied","Data":"82d11715c473edcc363a6e3392afc414dbf45b7d4125214c9869af36eb549af4"} Nov 24 13:51:49 crc kubenswrapper[4824]: I1124 13:51:49.957430 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:51:49 crc kubenswrapper[4824]: I1124 13:51:49.992942 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"18802c46-869d-4e00-848e-79df8665851d\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " Nov 24 13:51:49 crc kubenswrapper[4824]: I1124 13:51:49.993002 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-ssh-key\") pod \"18802c46-869d-4e00-848e-79df8665851d\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " Nov 24 13:51:49 crc kubenswrapper[4824]: I1124 13:51:49.993174 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-nova-metadata-neutron-config-0\") pod \"18802c46-869d-4e00-848e-79df8665851d\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " Nov 24 13:51:49 crc kubenswrapper[4824]: I1124 13:51:49.993220 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlv7c\" (UniqueName: \"kubernetes.io/projected/18802c46-869d-4e00-848e-79df8665851d-kube-api-access-zlv7c\") pod \"18802c46-869d-4e00-848e-79df8665851d\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " Nov 24 13:51:49 crc kubenswrapper[4824]: I1124 13:51:49.993248 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-inventory\") pod \"18802c46-869d-4e00-848e-79df8665851d\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " Nov 24 13:51:49 crc kubenswrapper[4824]: I1124 13:51:49.993305 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-metadata-combined-ca-bundle\") pod \"18802c46-869d-4e00-848e-79df8665851d\" (UID: \"18802c46-869d-4e00-848e-79df8665851d\") " Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.002653 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18802c46-869d-4e00-848e-79df8665851d-kube-api-access-zlv7c" (OuterVolumeSpecName: "kube-api-access-zlv7c") pod "18802c46-869d-4e00-848e-79df8665851d" (UID: "18802c46-869d-4e00-848e-79df8665851d"). InnerVolumeSpecName "kube-api-access-zlv7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.002779 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "18802c46-869d-4e00-848e-79df8665851d" (UID: "18802c46-869d-4e00-848e-79df8665851d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.034938 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "18802c46-869d-4e00-848e-79df8665851d" (UID: "18802c46-869d-4e00-848e-79df8665851d"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.035064 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-inventory" (OuterVolumeSpecName: "inventory") pod "18802c46-869d-4e00-848e-79df8665851d" (UID: "18802c46-869d-4e00-848e-79df8665851d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.041580 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18802c46-869d-4e00-848e-79df8665851d" (UID: "18802c46-869d-4e00-848e-79df8665851d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.053824 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "18802c46-869d-4e00-848e-79df8665851d" (UID: "18802c46-869d-4e00-848e-79df8665851d"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.096056 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlv7c\" (UniqueName: \"kubernetes.io/projected/18802c46-869d-4e00-848e-79df8665851d-kube-api-access-zlv7c\") on node \"crc\" DevicePath \"\"" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.096102 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.096118 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.096132 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.096146 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.096157 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18802c46-869d-4e00-848e-79df8665851d-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.516287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" event={"ID":"18802c46-869d-4e00-848e-79df8665851d","Type":"ContainerDied","Data":"234d83d9fd9d21e66bfc5f360aa4ade8822ecb034b52cceb394d39cc2428ccd2"} Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.516569 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="234d83d9fd9d21e66bfc5f360aa4ade8822ecb034b52cceb394d39cc2428ccd2" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.516346 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.616866 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76"] Nov 24 13:51:50 crc kubenswrapper[4824]: E1124 13:51:50.617442 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18802c46-869d-4e00-848e-79df8665851d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.617519 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="18802c46-869d-4e00-848e-79df8665851d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.617784 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="18802c46-869d-4e00-848e-79df8665851d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.618451 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.620623 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.621098 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.621450 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.622027 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.622901 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.637879 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76"] Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.705990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vcbv\" (UniqueName: \"kubernetes.io/projected/fa9782de-a426-4c52-a142-335c78e55a33-kube-api-access-4vcbv\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.706082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.706139 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.706192 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.706396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.808756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.808953 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.809057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vcbv\" (UniqueName: \"kubernetes.io/projected/fa9782de-a426-4c52-a142-335c78e55a33-kube-api-access-4vcbv\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.809142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.809216 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.815265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.815914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.816240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.817363 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.834054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vcbv\" (UniqueName: \"kubernetes.io/projected/fa9782de-a426-4c52-a142-335c78e55a33-kube-api-access-4vcbv\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-smm76\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:50 crc kubenswrapper[4824]: I1124 13:51:50.937734 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:51:51 crc kubenswrapper[4824]: I1124 13:51:51.532250 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76"] Nov 24 13:51:52 crc kubenswrapper[4824]: I1124 13:51:52.536055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" event={"ID":"fa9782de-a426-4c52-a142-335c78e55a33","Type":"ContainerStarted","Data":"c1e60bc6f957b25a37dc0795bc269ff0c4f9046763816a7246f2e293dcd5e6a6"} Nov 24 13:51:52 crc kubenswrapper[4824]: I1124 13:51:52.538511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" event={"ID":"fa9782de-a426-4c52-a142-335c78e55a33","Type":"ContainerStarted","Data":"5c7f791cb032a2fc50cecce4eea5af13664f5bdcd5cf07755b3f190cc5bb28e2"} Nov 24 13:51:52 crc kubenswrapper[4824]: I1124 13:51:52.564002 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" podStartSLOduration=2.383528982 podStartE2EDuration="2.563973653s" podCreationTimestamp="2025-11-24 13:51:50 +0000 UTC" firstStartedPulling="2025-11-24 13:51:51.558538717 +0000 UTC m=+2253.198078027" lastFinishedPulling="2025-11-24 13:51:51.738983388 +0000 UTC m=+2253.378522698" observedRunningTime="2025-11-24 13:51:52.55513265 +0000 UTC m=+2254.194671960" watchObservedRunningTime="2025-11-24 13:51:52.563973653 +0000 UTC m=+2254.203512963" Nov 24 13:52:42 crc kubenswrapper[4824]: I1124 13:52:42.843307 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2mkz5"] Nov 24 13:52:42 crc kubenswrapper[4824]: I1124 13:52:42.846447 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:42 crc kubenswrapper[4824]: I1124 13:52:42.867216 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2mkz5"] Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.037301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-utilities\") pod \"community-operators-2mkz5\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.037379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-catalog-content\") pod \"community-operators-2mkz5\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.037406 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4kqv\" (UniqueName: \"kubernetes.io/projected/14c8879d-00c4-4d7f-bf9a-c412de671742-kube-api-access-t4kqv\") pod \"community-operators-2mkz5\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.140501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-utilities\") pod \"community-operators-2mkz5\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.139893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-utilities\") pod \"community-operators-2mkz5\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.141946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-catalog-content\") pod \"community-operators-2mkz5\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.142308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-catalog-content\") pod \"community-operators-2mkz5\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.142377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4kqv\" (UniqueName: \"kubernetes.io/projected/14c8879d-00c4-4d7f-bf9a-c412de671742-kube-api-access-t4kqv\") pod \"community-operators-2mkz5\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.166799 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4kqv\" (UniqueName: \"kubernetes.io/projected/14c8879d-00c4-4d7f-bf9a-c412de671742-kube-api-access-t4kqv\") pod \"community-operators-2mkz5\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.172708 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:43 crc kubenswrapper[4824]: I1124 13:52:43.672452 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2mkz5"] Nov 24 13:52:44 crc kubenswrapper[4824]: I1124 13:52:44.034777 4824 generic.go:334] "Generic (PLEG): container finished" podID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerID="5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72" exitCode=0 Nov 24 13:52:44 crc kubenswrapper[4824]: I1124 13:52:44.034836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mkz5" event={"ID":"14c8879d-00c4-4d7f-bf9a-c412de671742","Type":"ContainerDied","Data":"5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72"} Nov 24 13:52:44 crc kubenswrapper[4824]: I1124 13:52:44.034861 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mkz5" event={"ID":"14c8879d-00c4-4d7f-bf9a-c412de671742","Type":"ContainerStarted","Data":"2175a52952e4fbb83db5b501be4bc3869cd44511a83ef3a65416deece7c4616a"} Nov 24 13:52:45 crc kubenswrapper[4824]: I1124 13:52:45.062615 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mkz5" event={"ID":"14c8879d-00c4-4d7f-bf9a-c412de671742","Type":"ContainerStarted","Data":"fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1"} Nov 24 13:52:46 crc kubenswrapper[4824]: I1124 13:52:46.071601 4824 generic.go:334] "Generic (PLEG): container finished" podID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerID="fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1" exitCode=0 Nov 24 13:52:46 crc kubenswrapper[4824]: I1124 13:52:46.071713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mkz5" event={"ID":"14c8879d-00c4-4d7f-bf9a-c412de671742","Type":"ContainerDied","Data":"fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1"} Nov 24 13:52:47 crc kubenswrapper[4824]: I1124 13:52:47.087209 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mkz5" event={"ID":"14c8879d-00c4-4d7f-bf9a-c412de671742","Type":"ContainerStarted","Data":"b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591"} Nov 24 13:52:47 crc kubenswrapper[4824]: I1124 13:52:47.119218 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2mkz5" podStartSLOduration=2.615033407 podStartE2EDuration="5.119199998s" podCreationTimestamp="2025-11-24 13:52:42 +0000 UTC" firstStartedPulling="2025-11-24 13:52:44.036260002 +0000 UTC m=+2305.675799312" lastFinishedPulling="2025-11-24 13:52:46.540426563 +0000 UTC m=+2308.179965903" observedRunningTime="2025-11-24 13:52:47.110267922 +0000 UTC m=+2308.749807262" watchObservedRunningTime="2025-11-24 13:52:47.119199998 +0000 UTC m=+2308.758739308" Nov 24 13:52:53 crc kubenswrapper[4824]: I1124 13:52:53.173517 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:53 crc kubenswrapper[4824]: I1124 13:52:53.174147 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:53 crc kubenswrapper[4824]: I1124 13:52:53.218194 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:54 crc kubenswrapper[4824]: I1124 13:52:54.246729 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:54 crc kubenswrapper[4824]: I1124 13:52:54.311581 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2mkz5"] Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.176273 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2mkz5" podUID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerName="registry-server" containerID="cri-o://b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591" gracePeriod=2 Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.701376 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.836620 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-utilities\") pod \"14c8879d-00c4-4d7f-bf9a-c412de671742\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.836793 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-catalog-content\") pod \"14c8879d-00c4-4d7f-bf9a-c412de671742\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.836868 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4kqv\" (UniqueName: \"kubernetes.io/projected/14c8879d-00c4-4d7f-bf9a-c412de671742-kube-api-access-t4kqv\") pod \"14c8879d-00c4-4d7f-bf9a-c412de671742\" (UID: \"14c8879d-00c4-4d7f-bf9a-c412de671742\") " Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.838119 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-utilities" (OuterVolumeSpecName: "utilities") pod "14c8879d-00c4-4d7f-bf9a-c412de671742" (UID: "14c8879d-00c4-4d7f-bf9a-c412de671742"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.842718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14c8879d-00c4-4d7f-bf9a-c412de671742-kube-api-access-t4kqv" (OuterVolumeSpecName: "kube-api-access-t4kqv") pod "14c8879d-00c4-4d7f-bf9a-c412de671742" (UID: "14c8879d-00c4-4d7f-bf9a-c412de671742"). InnerVolumeSpecName "kube-api-access-t4kqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.906288 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14c8879d-00c4-4d7f-bf9a-c412de671742" (UID: "14c8879d-00c4-4d7f-bf9a-c412de671742"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.939125 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.939158 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4kqv\" (UniqueName: \"kubernetes.io/projected/14c8879d-00c4-4d7f-bf9a-c412de671742-kube-api-access-t4kqv\") on node \"crc\" DevicePath \"\"" Nov 24 13:52:56 crc kubenswrapper[4824]: I1124 13:52:56.939170 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c8879d-00c4-4d7f-bf9a-c412de671742-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.189282 4824 generic.go:334] "Generic (PLEG): container finished" podID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerID="b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591" exitCode=0 Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.189735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mkz5" event={"ID":"14c8879d-00c4-4d7f-bf9a-c412de671742","Type":"ContainerDied","Data":"b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591"} Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.189770 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mkz5" event={"ID":"14c8879d-00c4-4d7f-bf9a-c412de671742","Type":"ContainerDied","Data":"2175a52952e4fbb83db5b501be4bc3869cd44511a83ef3a65416deece7c4616a"} Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.189792 4824 scope.go:117] "RemoveContainer" containerID="b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.189972 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2mkz5" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.233404 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2mkz5"] Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.245427 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2mkz5"] Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.266069 4824 scope.go:117] "RemoveContainer" containerID="fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.289788 4824 scope.go:117] "RemoveContainer" containerID="5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.332420 4824 scope.go:117] "RemoveContainer" containerID="b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591" Nov 24 13:52:57 crc kubenswrapper[4824]: E1124 13:52:57.332979 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591\": container with ID starting with b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591 not found: ID does not exist" containerID="b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.333011 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591"} err="failed to get container status \"b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591\": rpc error: code = NotFound desc = could not find container \"b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591\": container with ID starting with b6079ac35c36e9c1a644a087c84900832552775ceef97f8dd301c5ea7e0df591 not found: ID does not exist" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.333031 4824 scope.go:117] "RemoveContainer" containerID="fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1" Nov 24 13:52:57 crc kubenswrapper[4824]: E1124 13:52:57.333556 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1\": container with ID starting with fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1 not found: ID does not exist" containerID="fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.333610 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1"} err="failed to get container status \"fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1\": rpc error: code = NotFound desc = could not find container \"fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1\": container with ID starting with fc80420df6fcf5d6d22b98ff1c3779b740a514762fd65c43e033aa344db32bf1 not found: ID does not exist" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.333644 4824 scope.go:117] "RemoveContainer" containerID="5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72" Nov 24 13:52:57 crc kubenswrapper[4824]: E1124 13:52:57.334131 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72\": container with ID starting with 5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72 not found: ID does not exist" containerID="5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72" Nov 24 13:52:57 crc kubenswrapper[4824]: I1124 13:52:57.334156 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72"} err="failed to get container status \"5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72\": rpc error: code = NotFound desc = could not find container \"5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72\": container with ID starting with 5844bf8867eebab606f30beaded9e3a01bd3699bc7caf22ad1e0860aecfe9f72 not found: ID does not exist" Nov 24 13:52:59 crc kubenswrapper[4824]: I1124 13:52:59.025350 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14c8879d-00c4-4d7f-bf9a-c412de671742" path="/var/lib/kubelet/pods/14c8879d-00c4-4d7f-bf9a-c412de671742/volumes" Nov 24 13:53:40 crc kubenswrapper[4824]: I1124 13:53:40.788432 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:53:40 crc kubenswrapper[4824]: I1124 13:53:40.789153 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:54:10 crc kubenswrapper[4824]: I1124 13:54:10.787588 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:54:10 crc kubenswrapper[4824]: I1124 13:54:10.788194 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:54:40 crc kubenswrapper[4824]: I1124 13:54:40.787470 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:54:40 crc kubenswrapper[4824]: I1124 13:54:40.787931 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:54:40 crc kubenswrapper[4824]: I1124 13:54:40.787967 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 13:54:40 crc kubenswrapper[4824]: I1124 13:54:40.788612 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:54:40 crc kubenswrapper[4824]: I1124 13:54:40.788657 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" gracePeriod=600 Nov 24 13:54:40 crc kubenswrapper[4824]: E1124 13:54:40.990571 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:54:41 crc kubenswrapper[4824]: I1124 13:54:41.226318 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" exitCode=0 Nov 24 13:54:41 crc kubenswrapper[4824]: I1124 13:54:41.226377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b"} Nov 24 13:54:41 crc kubenswrapper[4824]: I1124 13:54:41.226430 4824 scope.go:117] "RemoveContainer" containerID="f7c32c7c1ab5c79ebcd1b5d0a45adaba8003211cdde83e67ef3cbc6580aa7fd2" Nov 24 13:54:41 crc kubenswrapper[4824]: I1124 13:54:41.227339 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:54:41 crc kubenswrapper[4824]: E1124 13:54:41.227721 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:54:53 crc kubenswrapper[4824]: I1124 13:54:53.010517 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:54:53 crc kubenswrapper[4824]: E1124 13:54:53.011284 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:55:08 crc kubenswrapper[4824]: I1124 13:55:08.010216 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:55:08 crc kubenswrapper[4824]: E1124 13:55:08.010979 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:55:23 crc kubenswrapper[4824]: I1124 13:55:23.010364 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:55:23 crc kubenswrapper[4824]: E1124 13:55:23.011160 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:55:34 crc kubenswrapper[4824]: I1124 13:55:34.009682 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:55:34 crc kubenswrapper[4824]: E1124 13:55:34.011532 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:55:48 crc kubenswrapper[4824]: I1124 13:55:48.009797 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:55:48 crc kubenswrapper[4824]: E1124 13:55:48.010624 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:56:01 crc kubenswrapper[4824]: I1124 13:56:01.010773 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:56:01 crc kubenswrapper[4824]: E1124 13:56:01.011773 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:56:02 crc kubenswrapper[4824]: I1124 13:56:02.043976 4824 generic.go:334] "Generic (PLEG): container finished" podID="fa9782de-a426-4c52-a142-335c78e55a33" containerID="c1e60bc6f957b25a37dc0795bc269ff0c4f9046763816a7246f2e293dcd5e6a6" exitCode=0 Nov 24 13:56:02 crc kubenswrapper[4824]: I1124 13:56:02.044067 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" event={"ID":"fa9782de-a426-4c52-a142-335c78e55a33","Type":"ContainerDied","Data":"c1e60bc6f957b25a37dc0795bc269ff0c4f9046763816a7246f2e293dcd5e6a6"} Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.538137 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.631278 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-combined-ca-bundle\") pod \"fa9782de-a426-4c52-a142-335c78e55a33\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.631627 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-ssh-key\") pod \"fa9782de-a426-4c52-a142-335c78e55a33\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.631771 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-inventory\") pod \"fa9782de-a426-4c52-a142-335c78e55a33\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.631963 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-secret-0\") pod \"fa9782de-a426-4c52-a142-335c78e55a33\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.632080 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vcbv\" (UniqueName: \"kubernetes.io/projected/fa9782de-a426-4c52-a142-335c78e55a33-kube-api-access-4vcbv\") pod \"fa9782de-a426-4c52-a142-335c78e55a33\" (UID: \"fa9782de-a426-4c52-a142-335c78e55a33\") " Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.636410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "fa9782de-a426-4c52-a142-335c78e55a33" (UID: "fa9782de-a426-4c52-a142-335c78e55a33"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.644022 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa9782de-a426-4c52-a142-335c78e55a33-kube-api-access-4vcbv" (OuterVolumeSpecName: "kube-api-access-4vcbv") pod "fa9782de-a426-4c52-a142-335c78e55a33" (UID: "fa9782de-a426-4c52-a142-335c78e55a33"). InnerVolumeSpecName "kube-api-access-4vcbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.669945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-inventory" (OuterVolumeSpecName: "inventory") pod "fa9782de-a426-4c52-a142-335c78e55a33" (UID: "fa9782de-a426-4c52-a142-335c78e55a33"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.693972 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "fa9782de-a426-4c52-a142-335c78e55a33" (UID: "fa9782de-a426-4c52-a142-335c78e55a33"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.694412 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fa9782de-a426-4c52-a142-335c78e55a33" (UID: "fa9782de-a426-4c52-a142-335c78e55a33"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.734967 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.735017 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vcbv\" (UniqueName: \"kubernetes.io/projected/fa9782de-a426-4c52-a142-335c78e55a33-kube-api-access-4vcbv\") on node \"crc\" DevicePath \"\"" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.735029 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.735038 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:56:03 crc kubenswrapper[4824]: I1124 13:56:03.735048 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa9782de-a426-4c52-a142-335c78e55a33-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.064116 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" event={"ID":"fa9782de-a426-4c52-a142-335c78e55a33","Type":"ContainerDied","Data":"5c7f791cb032a2fc50cecce4eea5af13664f5bdcd5cf07755b3f190cc5bb28e2"} Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.064412 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c7f791cb032a2fc50cecce4eea5af13664f5bdcd5cf07755b3f190cc5bb28e2" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.064167 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-smm76" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.242552 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4"] Nov 24 13:56:04 crc kubenswrapper[4824]: E1124 13:56:04.243037 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerName="extract-utilities" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.243054 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerName="extract-utilities" Nov 24 13:56:04 crc kubenswrapper[4824]: E1124 13:56:04.243079 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa9782de-a426-4c52-a142-335c78e55a33" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.243089 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa9782de-a426-4c52-a142-335c78e55a33" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 13:56:04 crc kubenswrapper[4824]: E1124 13:56:04.243118 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerName="extract-content" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.243125 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerName="extract-content" Nov 24 13:56:04 crc kubenswrapper[4824]: E1124 13:56:04.243141 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerName="registry-server" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.243147 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerName="registry-server" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.243340 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c8879d-00c4-4d7f-bf9a-c412de671742" containerName="registry-server" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.243362 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa9782de-a426-4c52-a142-335c78e55a33" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.261567 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4"] Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.261668 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.265380 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.265522 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.265736 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.265874 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.265959 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.265875 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.266028 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.468156 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.468291 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.468378 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.468426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.468470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.468569 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.468585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.468642 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6svxx\" (UniqueName: \"kubernetes.io/projected/4c37785a-1f6f-44c9-8673-b7f9f465682a-kube-api-access-6svxx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.468900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.570360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.570432 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.570479 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.570516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.570554 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.570610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.570639 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.570676 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6svxx\" (UniqueName: \"kubernetes.io/projected/4c37785a-1f6f-44c9-8673-b7f9f465682a-kube-api-access-6svxx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.570728 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.572372 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.575014 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.575514 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.576075 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.577014 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.577500 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.578179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.578578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.592018 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6svxx\" (UniqueName: \"kubernetes.io/projected/4c37785a-1f6f-44c9-8673-b7f9f465682a-kube-api-access-6svxx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtdx4\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:04 crc kubenswrapper[4824]: I1124 13:56:04.680092 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:56:05 crc kubenswrapper[4824]: I1124 13:56:05.261433 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4"] Nov 24 13:56:05 crc kubenswrapper[4824]: W1124 13:56:05.261903 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c37785a_1f6f_44c9_8673_b7f9f465682a.slice/crio-68d83e7b7f6189a0111f1a92c0c5f4ca7e68216a554188d70911cbadebea2fcf WatchSource:0}: Error finding container 68d83e7b7f6189a0111f1a92c0c5f4ca7e68216a554188d70911cbadebea2fcf: Status 404 returned error can't find the container with id 68d83e7b7f6189a0111f1a92c0c5f4ca7e68216a554188d70911cbadebea2fcf Nov 24 13:56:05 crc kubenswrapper[4824]: I1124 13:56:05.265581 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:56:06 crc kubenswrapper[4824]: I1124 13:56:06.106518 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" event={"ID":"4c37785a-1f6f-44c9-8673-b7f9f465682a","Type":"ContainerStarted","Data":"68d83e7b7f6189a0111f1a92c0c5f4ca7e68216a554188d70911cbadebea2fcf"} Nov 24 13:56:07 crc kubenswrapper[4824]: I1124 13:56:07.123629 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" event={"ID":"4c37785a-1f6f-44c9-8673-b7f9f465682a","Type":"ContainerStarted","Data":"000aee7a904a61c047307b4fe1b0a23b53de4d4625ff7fd470bbf694a001b323"} Nov 24 13:56:15 crc kubenswrapper[4824]: I1124 13:56:15.009939 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:56:15 crc kubenswrapper[4824]: E1124 13:56:15.010708 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:56:26 crc kubenswrapper[4824]: I1124 13:56:26.010763 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:56:26 crc kubenswrapper[4824]: E1124 13:56:26.011507 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:56:38 crc kubenswrapper[4824]: I1124 13:56:38.010483 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:56:38 crc kubenswrapper[4824]: E1124 13:56:38.011484 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:56:51 crc kubenswrapper[4824]: I1124 13:56:51.011221 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:56:51 crc kubenswrapper[4824]: E1124 13:56:51.012272 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:57:06 crc kubenswrapper[4824]: I1124 13:57:06.010099 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:57:06 crc kubenswrapper[4824]: E1124 13:57:06.011095 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:57:19 crc kubenswrapper[4824]: I1124 13:57:19.018003 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:57:19 crc kubenswrapper[4824]: E1124 13:57:19.019329 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:57:33 crc kubenswrapper[4824]: I1124 13:57:33.010151 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:57:33 crc kubenswrapper[4824]: E1124 13:57:33.011185 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:57:46 crc kubenswrapper[4824]: I1124 13:57:46.011524 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:57:46 crc kubenswrapper[4824]: E1124 13:57:46.012797 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:57:57 crc kubenswrapper[4824]: I1124 13:57:57.009914 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:57:57 crc kubenswrapper[4824]: E1124 13:57:57.010536 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.233147 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" podStartSLOduration=114.469010571 podStartE2EDuration="1m55.2331251s" podCreationTimestamp="2025-11-24 13:56:04 +0000 UTC" firstStartedPulling="2025-11-24 13:56:05.265327876 +0000 UTC m=+2506.904867196" lastFinishedPulling="2025-11-24 13:56:06.029442405 +0000 UTC m=+2507.668981725" observedRunningTime="2025-11-24 13:56:07.15556932 +0000 UTC m=+2508.795108640" watchObservedRunningTime="2025-11-24 13:57:59.2331251 +0000 UTC m=+2620.872664410" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.238868 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6t7g9"] Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.241617 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.258027 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6t7g9"] Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.442712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-catalog-content\") pod \"redhat-operators-6t7g9\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.442784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9tpx\" (UniqueName: \"kubernetes.io/projected/373d11f5-b249-4054-9885-e3a9a684315d-kube-api-access-x9tpx\") pod \"redhat-operators-6t7g9\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.442978 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-utilities\") pod \"redhat-operators-6t7g9\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.544489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-catalog-content\") pod \"redhat-operators-6t7g9\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.544542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9tpx\" (UniqueName: \"kubernetes.io/projected/373d11f5-b249-4054-9885-e3a9a684315d-kube-api-access-x9tpx\") pod \"redhat-operators-6t7g9\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.544602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-utilities\") pod \"redhat-operators-6t7g9\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.545428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-utilities\") pod \"redhat-operators-6t7g9\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.545700 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-catalog-content\") pod \"redhat-operators-6t7g9\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.567561 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9tpx\" (UniqueName: \"kubernetes.io/projected/373d11f5-b249-4054-9885-e3a9a684315d-kube-api-access-x9tpx\") pod \"redhat-operators-6t7g9\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.568147 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:57:59 crc kubenswrapper[4824]: I1124 13:57:59.903102 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6t7g9"] Nov 24 13:58:00 crc kubenswrapper[4824]: I1124 13:58:00.298828 4824 generic.go:334] "Generic (PLEG): container finished" podID="373d11f5-b249-4054-9885-e3a9a684315d" containerID="0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a" exitCode=0 Nov 24 13:58:00 crc kubenswrapper[4824]: I1124 13:58:00.298941 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6t7g9" event={"ID":"373d11f5-b249-4054-9885-e3a9a684315d","Type":"ContainerDied","Data":"0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a"} Nov 24 13:58:00 crc kubenswrapper[4824]: I1124 13:58:00.299186 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6t7g9" event={"ID":"373d11f5-b249-4054-9885-e3a9a684315d","Type":"ContainerStarted","Data":"f2ad860e8259a4b8298869360fb7be0d1dd34672f870b9a1459db5a21961f278"} Nov 24 13:58:09 crc kubenswrapper[4824]: I1124 13:58:09.019962 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:58:09 crc kubenswrapper[4824]: E1124 13:58:09.021085 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:58:10 crc kubenswrapper[4824]: I1124 13:58:10.388545 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6t7g9" event={"ID":"373d11f5-b249-4054-9885-e3a9a684315d","Type":"ContainerStarted","Data":"650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212"} Nov 24 13:58:12 crc kubenswrapper[4824]: I1124 13:58:12.408183 4824 generic.go:334] "Generic (PLEG): container finished" podID="373d11f5-b249-4054-9885-e3a9a684315d" containerID="650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212" exitCode=0 Nov 24 13:58:12 crc kubenswrapper[4824]: I1124 13:58:12.408251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6t7g9" event={"ID":"373d11f5-b249-4054-9885-e3a9a684315d","Type":"ContainerDied","Data":"650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212"} Nov 24 13:58:13 crc kubenswrapper[4824]: I1124 13:58:13.422931 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6t7g9" event={"ID":"373d11f5-b249-4054-9885-e3a9a684315d","Type":"ContainerStarted","Data":"160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a"} Nov 24 13:58:13 crc kubenswrapper[4824]: I1124 13:58:13.458108 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6t7g9" podStartSLOduration=1.848654649 podStartE2EDuration="14.458086571s" podCreationTimestamp="2025-11-24 13:57:59 +0000 UTC" firstStartedPulling="2025-11-24 13:58:00.300820374 +0000 UTC m=+2621.940359684" lastFinishedPulling="2025-11-24 13:58:12.910252296 +0000 UTC m=+2634.549791606" observedRunningTime="2025-11-24 13:58:13.448720308 +0000 UTC m=+2635.088259638" watchObservedRunningTime="2025-11-24 13:58:13.458086571 +0000 UTC m=+2635.097625881" Nov 24 13:58:19 crc kubenswrapper[4824]: I1124 13:58:19.569527 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:58:19 crc kubenswrapper[4824]: I1124 13:58:19.569993 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:58:19 crc kubenswrapper[4824]: I1124 13:58:19.612716 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:58:20 crc kubenswrapper[4824]: I1124 13:58:20.541997 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 13:58:20 crc kubenswrapper[4824]: I1124 13:58:20.623259 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6t7g9"] Nov 24 13:58:20 crc kubenswrapper[4824]: I1124 13:58:20.727052 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n8f7w"] Nov 24 13:58:20 crc kubenswrapper[4824]: I1124 13:58:20.727317 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n8f7w" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerName="registry-server" containerID="cri-o://6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8" gracePeriod=2 Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.248199 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.407887 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtdtt\" (UniqueName: \"kubernetes.io/projected/ecd285b6-a57e-4499-aa95-b3bd40d67be3-kube-api-access-vtdtt\") pod \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.408151 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-utilities\") pod \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.408504 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-catalog-content\") pod \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\" (UID: \"ecd285b6-a57e-4499-aa95-b3bd40d67be3\") " Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.409237 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-utilities" (OuterVolumeSpecName: "utilities") pod "ecd285b6-a57e-4499-aa95-b3bd40d67be3" (UID: "ecd285b6-a57e-4499-aa95-b3bd40d67be3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.418998 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd285b6-a57e-4499-aa95-b3bd40d67be3-kube-api-access-vtdtt" (OuterVolumeSpecName: "kube-api-access-vtdtt") pod "ecd285b6-a57e-4499-aa95-b3bd40d67be3" (UID: "ecd285b6-a57e-4499-aa95-b3bd40d67be3"). InnerVolumeSpecName "kube-api-access-vtdtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.499775 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ecd285b6-a57e-4499-aa95-b3bd40d67be3" (UID: "ecd285b6-a57e-4499-aa95-b3bd40d67be3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.508180 4824 generic.go:334] "Generic (PLEG): container finished" podID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerID="6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8" exitCode=0 Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.508227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8f7w" event={"ID":"ecd285b6-a57e-4499-aa95-b3bd40d67be3","Type":"ContainerDied","Data":"6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8"} Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.508245 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8f7w" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.508267 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8f7w" event={"ID":"ecd285b6-a57e-4499-aa95-b3bd40d67be3","Type":"ContainerDied","Data":"d8ba3adc683ab4f8c7237b30272e486cfea5933c510020bffd1fcfef681770dc"} Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.508284 4824 scope.go:117] "RemoveContainer" containerID="6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.510402 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtdtt\" (UniqueName: \"kubernetes.io/projected/ecd285b6-a57e-4499-aa95-b3bd40d67be3-kube-api-access-vtdtt\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.510500 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.510574 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd285b6-a57e-4499-aa95-b3bd40d67be3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.554977 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n8f7w"] Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.555030 4824 scope.go:117] "RemoveContainer" containerID="5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.558172 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n8f7w"] Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.585407 4824 scope.go:117] "RemoveContainer" containerID="c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.648864 4824 scope.go:117] "RemoveContainer" containerID="6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8" Nov 24 13:58:21 crc kubenswrapper[4824]: E1124 13:58:21.649294 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8\": container with ID starting with 6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8 not found: ID does not exist" containerID="6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.649330 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8"} err="failed to get container status \"6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8\": rpc error: code = NotFound desc = could not find container \"6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8\": container with ID starting with 6469459b58994cc69743f476b3336bd9c7db528fc5f99830819c6b6b83ac9ef8 not found: ID does not exist" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.649355 4824 scope.go:117] "RemoveContainer" containerID="5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1" Nov 24 13:58:21 crc kubenswrapper[4824]: E1124 13:58:21.651403 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1\": container with ID starting with 5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1 not found: ID does not exist" containerID="5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.651432 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1"} err="failed to get container status \"5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1\": rpc error: code = NotFound desc = could not find container \"5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1\": container with ID starting with 5edf194f66a3ef9e7eb244ad40dc0f8b2dea6249cf4c960551b1bde6885fa9c1 not found: ID does not exist" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.651452 4824 scope.go:117] "RemoveContainer" containerID="c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395" Nov 24 13:58:21 crc kubenswrapper[4824]: E1124 13:58:21.652307 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395\": container with ID starting with c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395 not found: ID does not exist" containerID="c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395" Nov 24 13:58:21 crc kubenswrapper[4824]: I1124 13:58:21.652333 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395"} err="failed to get container status \"c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395\": rpc error: code = NotFound desc = could not find container \"c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395\": container with ID starting with c006e13395b0cdfb9c2961275335e00d3bc29ec8175799112f9e33accfc9b395 not found: ID does not exist" Nov 24 13:58:22 crc kubenswrapper[4824]: I1124 13:58:22.420604 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-8cc59fd67-2j94h" podUID="57b08eb1-c1bf-4b56-880f-4d8adeaa32fc" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 24 13:58:23 crc kubenswrapper[4824]: I1124 13:58:23.018685 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" path="/var/lib/kubelet/pods/ecd285b6-a57e-4499-aa95-b3bd40d67be3/volumes" Nov 24 13:58:24 crc kubenswrapper[4824]: I1124 13:58:24.010503 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:58:24 crc kubenswrapper[4824]: E1124 13:58:24.010809 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:58:39 crc kubenswrapper[4824]: I1124 13:58:39.010330 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:58:39 crc kubenswrapper[4824]: E1124 13:58:39.011172 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:58:54 crc kubenswrapper[4824]: I1124 13:58:54.011197 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:58:54 crc kubenswrapper[4824]: E1124 13:58:54.012177 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:59:03 crc kubenswrapper[4824]: I1124 13:59:03.889183 4824 generic.go:334] "Generic (PLEG): container finished" podID="4c37785a-1f6f-44c9-8673-b7f9f465682a" containerID="000aee7a904a61c047307b4fe1b0a23b53de4d4625ff7fd470bbf694a001b323" exitCode=0 Nov 24 13:59:03 crc kubenswrapper[4824]: I1124 13:59:03.889283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" event={"ID":"4c37785a-1f6f-44c9-8673-b7f9f465682a","Type":"ContainerDied","Data":"000aee7a904a61c047307b4fe1b0a23b53de4d4625ff7fd470bbf694a001b323"} Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.294405 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.351702 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6svxx\" (UniqueName: \"kubernetes.io/projected/4c37785a-1f6f-44c9-8673-b7f9f465682a-kube-api-access-6svxx\") pod \"4c37785a-1f6f-44c9-8673-b7f9f465682a\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.351870 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-0\") pod \"4c37785a-1f6f-44c9-8673-b7f9f465682a\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.352015 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-extra-config-0\") pod \"4c37785a-1f6f-44c9-8673-b7f9f465682a\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.352088 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-ssh-key\") pod \"4c37785a-1f6f-44c9-8673-b7f9f465682a\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.352144 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-0\") pod \"4c37785a-1f6f-44c9-8673-b7f9f465682a\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.352204 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-1\") pod \"4c37785a-1f6f-44c9-8673-b7f9f465682a\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.352259 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-1\") pod \"4c37785a-1f6f-44c9-8673-b7f9f465682a\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.352330 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-combined-ca-bundle\") pod \"4c37785a-1f6f-44c9-8673-b7f9f465682a\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.352368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-inventory\") pod \"4c37785a-1f6f-44c9-8673-b7f9f465682a\" (UID: \"4c37785a-1f6f-44c9-8673-b7f9f465682a\") " Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.362328 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c37785a-1f6f-44c9-8673-b7f9f465682a-kube-api-access-6svxx" (OuterVolumeSpecName: "kube-api-access-6svxx") pod "4c37785a-1f6f-44c9-8673-b7f9f465682a" (UID: "4c37785a-1f6f-44c9-8673-b7f9f465682a"). InnerVolumeSpecName "kube-api-access-6svxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.378338 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4c37785a-1f6f-44c9-8673-b7f9f465682a" (UID: "4c37785a-1f6f-44c9-8673-b7f9f465682a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.380469 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4c37785a-1f6f-44c9-8673-b7f9f465682a" (UID: "4c37785a-1f6f-44c9-8673-b7f9f465682a"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.388260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4c37785a-1f6f-44c9-8673-b7f9f465682a" (UID: "4c37785a-1f6f-44c9-8673-b7f9f465682a"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.392945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4c37785a-1f6f-44c9-8673-b7f9f465682a" (UID: "4c37785a-1f6f-44c9-8673-b7f9f465682a"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.397259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4c37785a-1f6f-44c9-8673-b7f9f465682a" (UID: "4c37785a-1f6f-44c9-8673-b7f9f465682a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.398160 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4c37785a-1f6f-44c9-8673-b7f9f465682a" (UID: "4c37785a-1f6f-44c9-8673-b7f9f465682a"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.400212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-inventory" (OuterVolumeSpecName: "inventory") pod "4c37785a-1f6f-44c9-8673-b7f9f465682a" (UID: "4c37785a-1f6f-44c9-8673-b7f9f465682a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.404017 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4c37785a-1f6f-44c9-8673-b7f9f465682a" (UID: "4c37785a-1f6f-44c9-8673-b7f9f465682a"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.454642 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.454675 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.454686 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.454695 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.454705 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6svxx\" (UniqueName: \"kubernetes.io/projected/4c37785a-1f6f-44c9-8673-b7f9f465682a-kube-api-access-6svxx\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.454714 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.454724 4824 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.454732 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.454742 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4c37785a-1f6f-44c9-8673-b7f9f465682a-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.964052 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.964397 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtdx4" event={"ID":"4c37785a-1f6f-44c9-8673-b7f9f465682a","Type":"ContainerDied","Data":"68d83e7b7f6189a0111f1a92c0c5f4ca7e68216a554188d70911cbadebea2fcf"} Nov 24 13:59:05 crc kubenswrapper[4824]: I1124 13:59:05.964523 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68d83e7b7f6189a0111f1a92c0c5f4ca7e68216a554188d70911cbadebea2fcf" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.010844 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:59:06 crc kubenswrapper[4824]: E1124 13:59:06.011144 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.082256 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98"] Nov 24 13:59:06 crc kubenswrapper[4824]: E1124 13:59:06.082694 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerName="extract-content" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.082710 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerName="extract-content" Nov 24 13:59:06 crc kubenswrapper[4824]: E1124 13:59:06.082729 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c37785a-1f6f-44c9-8673-b7f9f465682a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.082742 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c37785a-1f6f-44c9-8673-b7f9f465682a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 13:59:06 crc kubenswrapper[4824]: E1124 13:59:06.082757 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerName="extract-utilities" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.082766 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerName="extract-utilities" Nov 24 13:59:06 crc kubenswrapper[4824]: E1124 13:59:06.082785 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerName="registry-server" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.082791 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerName="registry-server" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.083060 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c37785a-1f6f-44c9-8673-b7f9f465682a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.083077 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd285b6-a57e-4499-aa95-b3bd40d67be3" containerName="registry-server" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.083773 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.089850 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.089954 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.089991 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ch5l2" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.090079 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.090107 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.102054 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98"] Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.169327 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.169674 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjrq5\" (UniqueName: \"kubernetes.io/projected/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-kube-api-access-hjrq5\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.169841 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.170146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.170344 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.170444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.170528 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.272729 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjrq5\" (UniqueName: \"kubernetes.io/projected/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-kube-api-access-hjrq5\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.272935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.273112 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.273248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.273891 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.274117 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.274205 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.278961 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.278985 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.279573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.281125 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.281246 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.282515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.289730 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjrq5\" (UniqueName: \"kubernetes.io/projected/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-kube-api-access-hjrq5\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k6x98\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:06 crc kubenswrapper[4824]: I1124 13:59:06.404993 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 13:59:07 crc kubenswrapper[4824]: I1124 13:59:07.037466 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98"] Nov 24 13:59:07 crc kubenswrapper[4824]: I1124 13:59:07.986026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" event={"ID":"3fdb12da-712c-4ebe-97d1-4ebc46a1db80","Type":"ContainerStarted","Data":"fd168b91699f1d2cd1878fb6b837e2a42284da62a91ed7acd2698a214fe822a5"} Nov 24 13:59:07 crc kubenswrapper[4824]: I1124 13:59:07.986553 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" event={"ID":"3fdb12da-712c-4ebe-97d1-4ebc46a1db80","Type":"ContainerStarted","Data":"019e5749389a413afcff2f2e11fe991eb37beea8c542bd9eb8bb8a7db24954ee"} Nov 24 13:59:08 crc kubenswrapper[4824]: I1124 13:59:08.005777 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" podStartSLOduration=1.837696491 podStartE2EDuration="2.005757908s" podCreationTimestamp="2025-11-24 13:59:06 +0000 UTC" firstStartedPulling="2025-11-24 13:59:07.029629569 +0000 UTC m=+2688.669168919" lastFinishedPulling="2025-11-24 13:59:07.197691026 +0000 UTC m=+2688.837230336" observedRunningTime="2025-11-24 13:59:08.004353263 +0000 UTC m=+2689.643892613" watchObservedRunningTime="2025-11-24 13:59:08.005757908 +0000 UTC m=+2689.645297218" Nov 24 13:59:18 crc kubenswrapper[4824]: I1124 13:59:18.011370 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:59:18 crc kubenswrapper[4824]: E1124 13:59:18.012339 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:59:32 crc kubenswrapper[4824]: I1124 13:59:32.010259 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:59:32 crc kubenswrapper[4824]: E1124 13:59:32.011396 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 13:59:44 crc kubenswrapper[4824]: I1124 13:59:44.010944 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 13:59:44 crc kubenswrapper[4824]: I1124 13:59:44.345377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"9a8bedcb7448d68b9ae3979be0e7a2d503d2d1ac96679328d965b1f472005e8d"} Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.151453 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj"] Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.153607 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.156110 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.157028 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.161174 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj"] Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.282851 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-config-volume\") pod \"collect-profiles-29399880-gf6sj\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.282935 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-secret-volume\") pod \"collect-profiles-29399880-gf6sj\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.283150 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b64j\" (UniqueName: \"kubernetes.io/projected/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-kube-api-access-8b64j\") pod \"collect-profiles-29399880-gf6sj\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.384452 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b64j\" (UniqueName: \"kubernetes.io/projected/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-kube-api-access-8b64j\") pod \"collect-profiles-29399880-gf6sj\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.384539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-config-volume\") pod \"collect-profiles-29399880-gf6sj\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.384586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-secret-volume\") pod \"collect-profiles-29399880-gf6sj\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.386444 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-config-volume\") pod \"collect-profiles-29399880-gf6sj\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.390797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-secret-volume\") pod \"collect-profiles-29399880-gf6sj\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.409523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b64j\" (UniqueName: \"kubernetes.io/projected/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-kube-api-access-8b64j\") pod \"collect-profiles-29399880-gf6sj\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.477196 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:00 crc kubenswrapper[4824]: I1124 14:00:00.948564 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj"] Nov 24 14:00:01 crc kubenswrapper[4824]: I1124 14:00:01.575921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" event={"ID":"0e023e86-3ed5-431d-89ec-9e51e1d6ace7","Type":"ContainerStarted","Data":"0e70d7f6940258927458d61e620f67477c91fca6e9c3706299e945917b6cee8b"} Nov 24 14:00:01 crc kubenswrapper[4824]: I1124 14:00:01.576209 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" event={"ID":"0e023e86-3ed5-431d-89ec-9e51e1d6ace7","Type":"ContainerStarted","Data":"b6caf9df5faa6a60ab1462ae932745202036ab2850b2b1aed49d2491e9283f5b"} Nov 24 14:00:01 crc kubenswrapper[4824]: I1124 14:00:01.601276 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" podStartSLOduration=1.601260246 podStartE2EDuration="1.601260246s" podCreationTimestamp="2025-11-24 14:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:00:01.598745784 +0000 UTC m=+2743.238285114" watchObservedRunningTime="2025-11-24 14:00:01.601260246 +0000 UTC m=+2743.240799556" Nov 24 14:00:02 crc kubenswrapper[4824]: I1124 14:00:02.592640 4824 generic.go:334] "Generic (PLEG): container finished" podID="0e023e86-3ed5-431d-89ec-9e51e1d6ace7" containerID="0e70d7f6940258927458d61e620f67477c91fca6e9c3706299e945917b6cee8b" exitCode=0 Nov 24 14:00:02 crc kubenswrapper[4824]: I1124 14:00:02.592742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" event={"ID":"0e023e86-3ed5-431d-89ec-9e51e1d6ace7","Type":"ContainerDied","Data":"0e70d7f6940258927458d61e620f67477c91fca6e9c3706299e945917b6cee8b"} Nov 24 14:00:03 crc kubenswrapper[4824]: I1124 14:00:03.897715 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:03 crc kubenswrapper[4824]: I1124 14:00:03.968488 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-config-volume\") pod \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " Nov 24 14:00:03 crc kubenswrapper[4824]: I1124 14:00:03.969097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-secret-volume\") pod \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " Nov 24 14:00:03 crc kubenswrapper[4824]: I1124 14:00:03.969361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b64j\" (UniqueName: \"kubernetes.io/projected/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-kube-api-access-8b64j\") pod \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\" (UID: \"0e023e86-3ed5-431d-89ec-9e51e1d6ace7\") " Nov 24 14:00:03 crc kubenswrapper[4824]: I1124 14:00:03.969429 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-config-volume" (OuterVolumeSpecName: "config-volume") pod "0e023e86-3ed5-431d-89ec-9e51e1d6ace7" (UID: "0e023e86-3ed5-431d-89ec-9e51e1d6ace7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:00:03 crc kubenswrapper[4824]: I1124 14:00:03.970037 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:03 crc kubenswrapper[4824]: I1124 14:00:03.976285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-kube-api-access-8b64j" (OuterVolumeSpecName: "kube-api-access-8b64j") pod "0e023e86-3ed5-431d-89ec-9e51e1d6ace7" (UID: "0e023e86-3ed5-431d-89ec-9e51e1d6ace7"). InnerVolumeSpecName "kube-api-access-8b64j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:00:03 crc kubenswrapper[4824]: I1124 14:00:03.977212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0e023e86-3ed5-431d-89ec-9e51e1d6ace7" (UID: "0e023e86-3ed5-431d-89ec-9e51e1d6ace7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:00:04 crc kubenswrapper[4824]: I1124 14:00:04.070880 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b64j\" (UniqueName: \"kubernetes.io/projected/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-kube-api-access-8b64j\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:04 crc kubenswrapper[4824]: I1124 14:00:04.070941 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e023e86-3ed5-431d-89ec-9e51e1d6ace7-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:04 crc kubenswrapper[4824]: I1124 14:00:04.613377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" event={"ID":"0e023e86-3ed5-431d-89ec-9e51e1d6ace7","Type":"ContainerDied","Data":"b6caf9df5faa6a60ab1462ae932745202036ab2850b2b1aed49d2491e9283f5b"} Nov 24 14:00:04 crc kubenswrapper[4824]: I1124 14:00:04.613726 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6caf9df5faa6a60ab1462ae932745202036ab2850b2b1aed49d2491e9283f5b" Nov 24 14:00:04 crc kubenswrapper[4824]: I1124 14:00:04.613466 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-gf6sj" Nov 24 14:00:04 crc kubenswrapper[4824]: I1124 14:00:04.680330 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz"] Nov 24 14:00:04 crc kubenswrapper[4824]: I1124 14:00:04.689275 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399835-qbzcz"] Nov 24 14:00:05 crc kubenswrapper[4824]: I1124 14:00:05.021610 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2beaa442-848d-433a-841d-a461328b0376" path="/var/lib/kubelet/pods/2beaa442-848d-433a-841d-a461328b0376/volumes" Nov 24 14:00:24 crc kubenswrapper[4824]: I1124 14:00:24.627119 4824 scope.go:117] "RemoveContainer" containerID="7b0f9a9418a7de9a058b534f4c86a03a4901fee55b51612f0f2510078874c67c" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.147697 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29399881-7g6ln"] Nov 24 14:01:00 crc kubenswrapper[4824]: E1124 14:01:00.150704 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e023e86-3ed5-431d-89ec-9e51e1d6ace7" containerName="collect-profiles" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.150731 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e023e86-3ed5-431d-89ec-9e51e1d6ace7" containerName="collect-profiles" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.151041 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e023e86-3ed5-431d-89ec-9e51e1d6ace7" containerName="collect-profiles" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.152000 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.157722 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399881-7g6ln"] Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.202411 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-fernet-keys\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.202493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-combined-ca-bundle\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.202533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-config-data\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.202880 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv6xg\" (UniqueName: \"kubernetes.io/projected/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-kube-api-access-sv6xg\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.305170 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv6xg\" (UniqueName: \"kubernetes.io/projected/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-kube-api-access-sv6xg\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.305273 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-fernet-keys\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.305322 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-combined-ca-bundle\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.305347 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-config-data\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.312165 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-config-data\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.313496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-fernet-keys\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.316517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-combined-ca-bundle\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.324534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv6xg\" (UniqueName: \"kubernetes.io/projected/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-kube-api-access-sv6xg\") pod \"keystone-cron-29399881-7g6ln\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:00 crc kubenswrapper[4824]: I1124 14:01:00.479024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:01 crc kubenswrapper[4824]: W1124 14:01:01.025452 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd24ebeb5_8ce9_42cb_95c6_29162b65eb35.slice/crio-a5b5ec1ee5e95fecd1f597c60a7215c729b288d4fb23007f0f9e3b5f75c08ef4 WatchSource:0}: Error finding container a5b5ec1ee5e95fecd1f597c60a7215c729b288d4fb23007f0f9e3b5f75c08ef4: Status 404 returned error can't find the container with id a5b5ec1ee5e95fecd1f597c60a7215c729b288d4fb23007f0f9e3b5f75c08ef4 Nov 24 14:01:01 crc kubenswrapper[4824]: I1124 14:01:01.028128 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399881-7g6ln"] Nov 24 14:01:01 crc kubenswrapper[4824]: I1124 14:01:01.712739 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399881-7g6ln" event={"ID":"d24ebeb5-8ce9-42cb-95c6-29162b65eb35","Type":"ContainerStarted","Data":"f2c4a6a7b8d82bd90c44215c73b2a6bdc1a9e0f7c399fe13624ef3b2758970e4"} Nov 24 14:01:01 crc kubenswrapper[4824]: I1124 14:01:01.713046 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399881-7g6ln" event={"ID":"d24ebeb5-8ce9-42cb-95c6-29162b65eb35","Type":"ContainerStarted","Data":"a5b5ec1ee5e95fecd1f597c60a7215c729b288d4fb23007f0f9e3b5f75c08ef4"} Nov 24 14:01:04 crc kubenswrapper[4824]: I1124 14:01:04.745002 4824 generic.go:334] "Generic (PLEG): container finished" podID="d24ebeb5-8ce9-42cb-95c6-29162b65eb35" containerID="f2c4a6a7b8d82bd90c44215c73b2a6bdc1a9e0f7c399fe13624ef3b2758970e4" exitCode=0 Nov 24 14:01:04 crc kubenswrapper[4824]: I1124 14:01:04.745483 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399881-7g6ln" event={"ID":"d24ebeb5-8ce9-42cb-95c6-29162b65eb35","Type":"ContainerDied","Data":"f2c4a6a7b8d82bd90c44215c73b2a6bdc1a9e0f7c399fe13624ef3b2758970e4"} Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.107782 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.228128 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv6xg\" (UniqueName: \"kubernetes.io/projected/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-kube-api-access-sv6xg\") pod \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.228316 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-combined-ca-bundle\") pod \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.228354 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-config-data\") pod \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.228436 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-fernet-keys\") pod \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\" (UID: \"d24ebeb5-8ce9-42cb-95c6-29162b65eb35\") " Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.233898 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d24ebeb5-8ce9-42cb-95c6-29162b65eb35" (UID: "d24ebeb5-8ce9-42cb-95c6-29162b65eb35"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.234026 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-kube-api-access-sv6xg" (OuterVolumeSpecName: "kube-api-access-sv6xg") pod "d24ebeb5-8ce9-42cb-95c6-29162b65eb35" (UID: "d24ebeb5-8ce9-42cb-95c6-29162b65eb35"). InnerVolumeSpecName "kube-api-access-sv6xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.262989 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d24ebeb5-8ce9-42cb-95c6-29162b65eb35" (UID: "d24ebeb5-8ce9-42cb-95c6-29162b65eb35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.300527 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-config-data" (OuterVolumeSpecName: "config-data") pod "d24ebeb5-8ce9-42cb-95c6-29162b65eb35" (UID: "d24ebeb5-8ce9-42cb-95c6-29162b65eb35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.330779 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.330846 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.330857 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.330866 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv6xg\" (UniqueName: \"kubernetes.io/projected/d24ebeb5-8ce9-42cb-95c6-29162b65eb35-kube-api-access-sv6xg\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.767484 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399881-7g6ln" event={"ID":"d24ebeb5-8ce9-42cb-95c6-29162b65eb35","Type":"ContainerDied","Data":"a5b5ec1ee5e95fecd1f597c60a7215c729b288d4fb23007f0f9e3b5f75c08ef4"} Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.767547 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5b5ec1ee5e95fecd1f597c60a7215c729b288d4fb23007f0f9e3b5f75c08ef4" Nov 24 14:01:06 crc kubenswrapper[4824]: I1124 14:01:06.767896 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399881-7g6ln" Nov 24 14:02:10 crc kubenswrapper[4824]: I1124 14:02:10.788337 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:02:10 crc kubenswrapper[4824]: I1124 14:02:10.788941 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:02:16 crc kubenswrapper[4824]: I1124 14:02:16.470870 4824 generic.go:334] "Generic (PLEG): container finished" podID="3fdb12da-712c-4ebe-97d1-4ebc46a1db80" containerID="fd168b91699f1d2cd1878fb6b837e2a42284da62a91ed7acd2698a214fe822a5" exitCode=0 Nov 24 14:02:16 crc kubenswrapper[4824]: I1124 14:02:16.470934 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" event={"ID":"3fdb12da-712c-4ebe-97d1-4ebc46a1db80","Type":"ContainerDied","Data":"fd168b91699f1d2cd1878fb6b837e2a42284da62a91ed7acd2698a214fe822a5"} Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.881038 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.893605 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-0\") pod \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.893658 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-inventory\") pod \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.893757 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-telemetry-combined-ca-bundle\") pod \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.893778 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-1\") pod \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.893894 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ssh-key\") pod \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.893976 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjrq5\" (UniqueName: \"kubernetes.io/projected/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-kube-api-access-hjrq5\") pod \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.894037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-2\") pod \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\" (UID: \"3fdb12da-712c-4ebe-97d1-4ebc46a1db80\") " Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.912043 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-kube-api-access-hjrq5" (OuterVolumeSpecName: "kube-api-access-hjrq5") pod "3fdb12da-712c-4ebe-97d1-4ebc46a1db80" (UID: "3fdb12da-712c-4ebe-97d1-4ebc46a1db80"). InnerVolumeSpecName "kube-api-access-hjrq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.928304 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-inventory" (OuterVolumeSpecName: "inventory") pod "3fdb12da-712c-4ebe-97d1-4ebc46a1db80" (UID: "3fdb12da-712c-4ebe-97d1-4ebc46a1db80"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.937910 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3fdb12da-712c-4ebe-97d1-4ebc46a1db80" (UID: "3fdb12da-712c-4ebe-97d1-4ebc46a1db80"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.940836 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "3fdb12da-712c-4ebe-97d1-4ebc46a1db80" (UID: "3fdb12da-712c-4ebe-97d1-4ebc46a1db80"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.942766 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3fdb12da-712c-4ebe-97d1-4ebc46a1db80" (UID: "3fdb12da-712c-4ebe-97d1-4ebc46a1db80"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.958675 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "3fdb12da-712c-4ebe-97d1-4ebc46a1db80" (UID: "3fdb12da-712c-4ebe-97d1-4ebc46a1db80"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.969435 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "3fdb12da-712c-4ebe-97d1-4ebc46a1db80" (UID: "3fdb12da-712c-4ebe-97d1-4ebc46a1db80"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.997348 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjrq5\" (UniqueName: \"kubernetes.io/projected/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-kube-api-access-hjrq5\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.997466 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.997528 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.997593 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.997655 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.997901 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:17 crc kubenswrapper[4824]: I1124 14:02:17.997980 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fdb12da-712c-4ebe-97d1-4ebc46a1db80-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4824]: I1124 14:02:18.526887 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" event={"ID":"3fdb12da-712c-4ebe-97d1-4ebc46a1db80","Type":"ContainerDied","Data":"019e5749389a413afcff2f2e11fe991eb37beea8c542bd9eb8bb8a7db24954ee"} Nov 24 14:02:18 crc kubenswrapper[4824]: I1124 14:02:18.526937 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="019e5749389a413afcff2f2e11fe991eb37beea8c542bd9eb8bb8a7db24954ee" Nov 24 14:02:18 crc kubenswrapper[4824]: I1124 14:02:18.527063 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k6x98" Nov 24 14:02:40 crc kubenswrapper[4824]: I1124 14:02:40.788375 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:02:40 crc kubenswrapper[4824]: I1124 14:02:40.789074 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:02:42 crc kubenswrapper[4824]: E1124 14:02:42.828178 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.188:33902->38.129.56.188:33927: write tcp 38.129.56.188:33902->38.129.56.188:33927: write: broken pipe Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.494658 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 14:03:01 crc kubenswrapper[4824]: E1124 14:03:01.495849 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fdb12da-712c-4ebe-97d1-4ebc46a1db80" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.495872 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fdb12da-712c-4ebe-97d1-4ebc46a1db80" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 14:03:01 crc kubenswrapper[4824]: E1124 14:03:01.495916 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d24ebeb5-8ce9-42cb-95c6-29162b65eb35" containerName="keystone-cron" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.495927 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d24ebeb5-8ce9-42cb-95c6-29162b65eb35" containerName="keystone-cron" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.496282 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fdb12da-712c-4ebe-97d1-4ebc46a1db80" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.496320 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d24ebeb5-8ce9-42cb-95c6-29162b65eb35" containerName="keystone-cron" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.497286 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.499583 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.500021 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qvsng" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.502406 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.506227 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.507950 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.637718 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.637778 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.637802 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.637846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-config-data\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.637881 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.637975 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx7jt\" (UniqueName: \"kubernetes.io/projected/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-kube-api-access-kx7jt\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.638029 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.638091 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.638181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.739246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.739300 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.739359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.739449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.739483 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.739500 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.739521 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-config-data\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.739547 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.739570 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx7jt\" (UniqueName: \"kubernetes.io/projected/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-kube-api-access-kx7jt\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.740278 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.741800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.742882 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.743336 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.743839 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-config-data\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.747138 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.747493 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.750047 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.761214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx7jt\" (UniqueName: \"kubernetes.io/projected/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-kube-api-access-kx7jt\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.780197 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " pod="openstack/tempest-tests-tempest" Nov 24 14:03:01 crc kubenswrapper[4824]: I1124 14:03:01.821000 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 14:03:02 crc kubenswrapper[4824]: I1124 14:03:02.279535 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 14:03:02 crc kubenswrapper[4824]: I1124 14:03:02.297039 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:03:02 crc kubenswrapper[4824]: I1124 14:03:02.945769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0","Type":"ContainerStarted","Data":"1b89ff540eff077c43467e51a8dfff04ade825414083d67c4890eabb89ce956b"} Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.039362 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tc5fb"] Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.042777 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.052166 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tc5fb"] Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.208036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hm7w\" (UniqueName: \"kubernetes.io/projected/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-kube-api-access-4hm7w\") pod \"community-operators-tc5fb\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.208888 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-catalog-content\") pod \"community-operators-tc5fb\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.208997 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-utilities\") pod \"community-operators-tc5fb\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.310407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-utilities\") pod \"community-operators-tc5fb\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.310501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hm7w\" (UniqueName: \"kubernetes.io/projected/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-kube-api-access-4hm7w\") pod \"community-operators-tc5fb\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.310553 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-catalog-content\") pod \"community-operators-tc5fb\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.310919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-utilities\") pod \"community-operators-tc5fb\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.310984 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-catalog-content\") pod \"community-operators-tc5fb\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.341222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hm7w\" (UniqueName: \"kubernetes.io/projected/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-kube-api-access-4hm7w\") pod \"community-operators-tc5fb\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:09 crc kubenswrapper[4824]: I1124 14:03:09.450424 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:10 crc kubenswrapper[4824]: I1124 14:03:10.035094 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tc5fb"] Nov 24 14:03:10 crc kubenswrapper[4824]: W1124 14:03:10.036799 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf996432_3c6f_4aaa_a216_2e1bc56d8bc0.slice/crio-74efa369006a17168703574eec6547463f432f6587dafce581201d461f0af3c7 WatchSource:0}: Error finding container 74efa369006a17168703574eec6547463f432f6587dafce581201d461f0af3c7: Status 404 returned error can't find the container with id 74efa369006a17168703574eec6547463f432f6587dafce581201d461f0af3c7 Nov 24 14:03:10 crc kubenswrapper[4824]: I1124 14:03:10.788775 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:03:10 crc kubenswrapper[4824]: I1124 14:03:10.788879 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:03:10 crc kubenswrapper[4824]: I1124 14:03:10.788951 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 14:03:10 crc kubenswrapper[4824]: I1124 14:03:10.789943 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9a8bedcb7448d68b9ae3979be0e7a2d503d2d1ac96679328d965b1f472005e8d"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:03:10 crc kubenswrapper[4824]: I1124 14:03:10.790004 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://9a8bedcb7448d68b9ae3979be0e7a2d503d2d1ac96679328d965b1f472005e8d" gracePeriod=600 Nov 24 14:03:11 crc kubenswrapper[4824]: I1124 14:03:11.030125 4824 generic.go:334] "Generic (PLEG): container finished" podID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerID="7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17" exitCode=0 Nov 24 14:03:11 crc kubenswrapper[4824]: I1124 14:03:11.030570 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tc5fb" event={"ID":"af996432-3c6f-4aaa-a216-2e1bc56d8bc0","Type":"ContainerDied","Data":"7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17"} Nov 24 14:03:11 crc kubenswrapper[4824]: I1124 14:03:11.030603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tc5fb" event={"ID":"af996432-3c6f-4aaa-a216-2e1bc56d8bc0","Type":"ContainerStarted","Data":"74efa369006a17168703574eec6547463f432f6587dafce581201d461f0af3c7"} Nov 24 14:03:11 crc kubenswrapper[4824]: I1124 14:03:11.035699 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="9a8bedcb7448d68b9ae3979be0e7a2d503d2d1ac96679328d965b1f472005e8d" exitCode=0 Nov 24 14:03:11 crc kubenswrapper[4824]: I1124 14:03:11.035755 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"9a8bedcb7448d68b9ae3979be0e7a2d503d2d1ac96679328d965b1f472005e8d"} Nov 24 14:03:11 crc kubenswrapper[4824]: I1124 14:03:11.035794 4824 scope.go:117] "RemoveContainer" containerID="cf8d1e67b9893682f76b73d157e67fe5c95961032cdb699140f1c1895a885e8b" Nov 24 14:03:12 crc kubenswrapper[4824]: I1124 14:03:12.046335 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c"} Nov 24 14:03:12 crc kubenswrapper[4824]: I1124 14:03:12.056666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tc5fb" event={"ID":"af996432-3c6f-4aaa-a216-2e1bc56d8bc0","Type":"ContainerStarted","Data":"2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f"} Nov 24 14:03:14 crc kubenswrapper[4824]: I1124 14:03:14.078783 4824 generic.go:334] "Generic (PLEG): container finished" podID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerID="2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f" exitCode=0 Nov 24 14:03:14 crc kubenswrapper[4824]: I1124 14:03:14.079027 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tc5fb" event={"ID":"af996432-3c6f-4aaa-a216-2e1bc56d8bc0","Type":"ContainerDied","Data":"2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f"} Nov 24 14:03:19 crc kubenswrapper[4824]: I1124 14:03:19.816878 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ztmfn"] Nov 24 14:03:19 crc kubenswrapper[4824]: I1124 14:03:19.819366 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:19 crc kubenswrapper[4824]: I1124 14:03:19.834304 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ztmfn"] Nov 24 14:03:19 crc kubenswrapper[4824]: I1124 14:03:19.966645 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-catalog-content\") pod \"certified-operators-ztmfn\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:19 crc kubenswrapper[4824]: I1124 14:03:19.966734 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jd6t\" (UniqueName: \"kubernetes.io/projected/bfb91335-3f3a-45da-86c0-4decb3bd5191-kube-api-access-7jd6t\") pod \"certified-operators-ztmfn\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:19 crc kubenswrapper[4824]: I1124 14:03:19.966874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-utilities\") pod \"certified-operators-ztmfn\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.068906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-utilities\") pod \"certified-operators-ztmfn\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.069247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-catalog-content\") pod \"certified-operators-ztmfn\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.069315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jd6t\" (UniqueName: \"kubernetes.io/projected/bfb91335-3f3a-45da-86c0-4decb3bd5191-kube-api-access-7jd6t\") pod \"certified-operators-ztmfn\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.070054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-catalog-content\") pod \"certified-operators-ztmfn\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.071394 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-utilities\") pod \"certified-operators-ztmfn\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.098059 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jd6t\" (UniqueName: \"kubernetes.io/projected/bfb91335-3f3a-45da-86c0-4decb3bd5191-kube-api-access-7jd6t\") pod \"certified-operators-ztmfn\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.141236 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.615119 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h8s6m"] Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.617395 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.630238 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h8s6m"] Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.706932 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-utilities\") pod \"redhat-marketplace-h8s6m\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.707023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w568b\" (UniqueName: \"kubernetes.io/projected/54e5f8c6-26f1-4318-b507-d71defd8ee8f-kube-api-access-w568b\") pod \"redhat-marketplace-h8s6m\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.707052 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-catalog-content\") pod \"redhat-marketplace-h8s6m\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.809107 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-catalog-content\") pod \"redhat-marketplace-h8s6m\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.809278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-utilities\") pod \"redhat-marketplace-h8s6m\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.809321 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w568b\" (UniqueName: \"kubernetes.io/projected/54e5f8c6-26f1-4318-b507-d71defd8ee8f-kube-api-access-w568b\") pod \"redhat-marketplace-h8s6m\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.809699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-utilities\") pod \"redhat-marketplace-h8s6m\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.811353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-catalog-content\") pod \"redhat-marketplace-h8s6m\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.829937 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w568b\" (UniqueName: \"kubernetes.io/projected/54e5f8c6-26f1-4318-b507-d71defd8ee8f-kube-api-access-w568b\") pod \"redhat-marketplace-h8s6m\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:20 crc kubenswrapper[4824]: I1124 14:03:20.950112 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:35 crc kubenswrapper[4824]: E1124 14:03:35.702266 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 24 14:03:35 crc kubenswrapper[4824]: E1124 14:03:35.702771 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kx7jt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:03:35 crc kubenswrapper[4824]: E1124 14:03:35.705364 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" Nov 24 14:03:36 crc kubenswrapper[4824]: I1124 14:03:36.109850 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h8s6m"] Nov 24 14:03:36 crc kubenswrapper[4824]: I1124 14:03:36.241600 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ztmfn"] Nov 24 14:03:36 crc kubenswrapper[4824]: I1124 14:03:36.294502 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ztmfn" event={"ID":"bfb91335-3f3a-45da-86c0-4decb3bd5191","Type":"ContainerStarted","Data":"368e1d546ae8ef4fa4478c88c223927169ad285243efb5c397351f9f278cbaff"} Nov 24 14:03:36 crc kubenswrapper[4824]: I1124 14:03:36.329215 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h8s6m" event={"ID":"54e5f8c6-26f1-4318-b507-d71defd8ee8f","Type":"ContainerStarted","Data":"c619578116dd74f567a1f24b7a0fd8639ca89cf7f809f343bb6e52ae31296f1c"} Nov 24 14:03:36 crc kubenswrapper[4824]: E1124 14:03:36.347117 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" Nov 24 14:03:37 crc kubenswrapper[4824]: I1124 14:03:37.343642 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tc5fb" event={"ID":"af996432-3c6f-4aaa-a216-2e1bc56d8bc0","Type":"ContainerStarted","Data":"2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290"} Nov 24 14:03:37 crc kubenswrapper[4824]: I1124 14:03:37.347431 4824 generic.go:334] "Generic (PLEG): container finished" podID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerID="7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6" exitCode=0 Nov 24 14:03:37 crc kubenswrapper[4824]: I1124 14:03:37.347502 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h8s6m" event={"ID":"54e5f8c6-26f1-4318-b507-d71defd8ee8f","Type":"ContainerDied","Data":"7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6"} Nov 24 14:03:37 crc kubenswrapper[4824]: I1124 14:03:37.350802 4824 generic.go:334] "Generic (PLEG): container finished" podID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerID="be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc" exitCode=0 Nov 24 14:03:37 crc kubenswrapper[4824]: I1124 14:03:37.350849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ztmfn" event={"ID":"bfb91335-3f3a-45da-86c0-4decb3bd5191","Type":"ContainerDied","Data":"be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc"} Nov 24 14:03:37 crc kubenswrapper[4824]: I1124 14:03:37.368196 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tc5fb" podStartSLOduration=3.197519693 podStartE2EDuration="28.368171425s" podCreationTimestamp="2025-11-24 14:03:09 +0000 UTC" firstStartedPulling="2025-11-24 14:03:11.031901276 +0000 UTC m=+2932.671440586" lastFinishedPulling="2025-11-24 14:03:36.202553008 +0000 UTC m=+2957.842092318" observedRunningTime="2025-11-24 14:03:37.366178446 +0000 UTC m=+2959.005717766" watchObservedRunningTime="2025-11-24 14:03:37.368171425 +0000 UTC m=+2959.007710745" Nov 24 14:03:38 crc kubenswrapper[4824]: I1124 14:03:38.361132 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ztmfn" event={"ID":"bfb91335-3f3a-45da-86c0-4decb3bd5191","Type":"ContainerStarted","Data":"571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b"} Nov 24 14:03:39 crc kubenswrapper[4824]: I1124 14:03:39.369996 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h8s6m" event={"ID":"54e5f8c6-26f1-4318-b507-d71defd8ee8f","Type":"ContainerStarted","Data":"390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9"} Nov 24 14:03:39 crc kubenswrapper[4824]: I1124 14:03:39.450734 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:39 crc kubenswrapper[4824]: I1124 14:03:39.450789 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:40 crc kubenswrapper[4824]: I1124 14:03:40.381488 4824 generic.go:334] "Generic (PLEG): container finished" podID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerID="390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9" exitCode=0 Nov 24 14:03:40 crc kubenswrapper[4824]: I1124 14:03:40.381541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h8s6m" event={"ID":"54e5f8c6-26f1-4318-b507-d71defd8ee8f","Type":"ContainerDied","Data":"390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9"} Nov 24 14:03:40 crc kubenswrapper[4824]: I1124 14:03:40.516928 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tc5fb" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="registry-server" probeResult="failure" output=< Nov 24 14:03:40 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:03:40 crc kubenswrapper[4824]: > Nov 24 14:03:42 crc kubenswrapper[4824]: I1124 14:03:42.403081 4824 generic.go:334] "Generic (PLEG): container finished" podID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerID="571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b" exitCode=0 Nov 24 14:03:42 crc kubenswrapper[4824]: I1124 14:03:42.403173 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ztmfn" event={"ID":"bfb91335-3f3a-45da-86c0-4decb3bd5191","Type":"ContainerDied","Data":"571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b"} Nov 24 14:03:42 crc kubenswrapper[4824]: I1124 14:03:42.406626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h8s6m" event={"ID":"54e5f8c6-26f1-4318-b507-d71defd8ee8f","Type":"ContainerStarted","Data":"4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de"} Nov 24 14:03:42 crc kubenswrapper[4824]: I1124 14:03:42.447922 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h8s6m" podStartSLOduration=18.241939951 podStartE2EDuration="22.447907226s" podCreationTimestamp="2025-11-24 14:03:20 +0000 UTC" firstStartedPulling="2025-11-24 14:03:37.34995717 +0000 UTC m=+2958.989496480" lastFinishedPulling="2025-11-24 14:03:41.555924445 +0000 UTC m=+2963.195463755" observedRunningTime="2025-11-24 14:03:42.443982618 +0000 UTC m=+2964.083521928" watchObservedRunningTime="2025-11-24 14:03:42.447907226 +0000 UTC m=+2964.087446536" Nov 24 14:03:43 crc kubenswrapper[4824]: I1124 14:03:43.421547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ztmfn" event={"ID":"bfb91335-3f3a-45da-86c0-4decb3bd5191","Type":"ContainerStarted","Data":"bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c"} Nov 24 14:03:43 crc kubenswrapper[4824]: I1124 14:03:43.445067 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ztmfn" podStartSLOduration=18.977889277 podStartE2EDuration="24.445044029s" podCreationTimestamp="2025-11-24 14:03:19 +0000 UTC" firstStartedPulling="2025-11-24 14:03:37.35238366 +0000 UTC m=+2958.991922990" lastFinishedPulling="2025-11-24 14:03:42.819538432 +0000 UTC m=+2964.459077742" observedRunningTime="2025-11-24 14:03:43.444503955 +0000 UTC m=+2965.084043265" watchObservedRunningTime="2025-11-24 14:03:43.445044029 +0000 UTC m=+2965.084583339" Nov 24 14:03:48 crc kubenswrapper[4824]: I1124 14:03:48.486144 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 14:03:50 crc kubenswrapper[4824]: I1124 14:03:50.141947 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:50 crc kubenswrapper[4824]: I1124 14:03:50.142406 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:50 crc kubenswrapper[4824]: I1124 14:03:50.188108 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:50 crc kubenswrapper[4824]: I1124 14:03:50.484164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0","Type":"ContainerStarted","Data":"bcd1546a48cc865529bb47c2eaa60d9ed935f25e27bade958e6f2482a2ccf4f0"} Nov 24 14:03:50 crc kubenswrapper[4824]: I1124 14:03:50.500603 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.313616867 podStartE2EDuration="50.500585924s" podCreationTimestamp="2025-11-24 14:03:00 +0000 UTC" firstStartedPulling="2025-11-24 14:03:02.296783649 +0000 UTC m=+2923.936322959" lastFinishedPulling="2025-11-24 14:03:48.483752706 +0000 UTC m=+2970.123292016" observedRunningTime="2025-11-24 14:03:50.497733763 +0000 UTC m=+2972.137273073" watchObservedRunningTime="2025-11-24 14:03:50.500585924 +0000 UTC m=+2972.140125234" Nov 24 14:03:50 crc kubenswrapper[4824]: I1124 14:03:50.505786 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tc5fb" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="registry-server" probeResult="failure" output=< Nov 24 14:03:50 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:03:50 crc kubenswrapper[4824]: > Nov 24 14:03:50 crc kubenswrapper[4824]: I1124 14:03:50.531578 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:50 crc kubenswrapper[4824]: I1124 14:03:50.951009 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:50 crc kubenswrapper[4824]: I1124 14:03:50.951479 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:51 crc kubenswrapper[4824]: I1124 14:03:51.026882 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:51 crc kubenswrapper[4824]: I1124 14:03:51.409466 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ztmfn"] Nov 24 14:03:51 crc kubenswrapper[4824]: I1124 14:03:51.541054 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:52 crc kubenswrapper[4824]: I1124 14:03:52.502205 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ztmfn" podUID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerName="registry-server" containerID="cri-o://bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c" gracePeriod=2 Nov 24 14:03:52 crc kubenswrapper[4824]: I1124 14:03:52.813349 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h8s6m"] Nov 24 14:03:52 crc kubenswrapper[4824]: I1124 14:03:52.965007 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.064027 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-utilities\") pod \"bfb91335-3f3a-45da-86c0-4decb3bd5191\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.064151 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jd6t\" (UniqueName: \"kubernetes.io/projected/bfb91335-3f3a-45da-86c0-4decb3bd5191-kube-api-access-7jd6t\") pod \"bfb91335-3f3a-45da-86c0-4decb3bd5191\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.064250 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-catalog-content\") pod \"bfb91335-3f3a-45da-86c0-4decb3bd5191\" (UID: \"bfb91335-3f3a-45da-86c0-4decb3bd5191\") " Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.068494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-utilities" (OuterVolumeSpecName: "utilities") pod "bfb91335-3f3a-45da-86c0-4decb3bd5191" (UID: "bfb91335-3f3a-45da-86c0-4decb3bd5191"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.075064 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb91335-3f3a-45da-86c0-4decb3bd5191-kube-api-access-7jd6t" (OuterVolumeSpecName: "kube-api-access-7jd6t") pod "bfb91335-3f3a-45da-86c0-4decb3bd5191" (UID: "bfb91335-3f3a-45da-86c0-4decb3bd5191"). InnerVolumeSpecName "kube-api-access-7jd6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.119634 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfb91335-3f3a-45da-86c0-4decb3bd5191" (UID: "bfb91335-3f3a-45da-86c0-4decb3bd5191"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.166299 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.166325 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jd6t\" (UniqueName: \"kubernetes.io/projected/bfb91335-3f3a-45da-86c0-4decb3bd5191-kube-api-access-7jd6t\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.166335 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb91335-3f3a-45da-86c0-4decb3bd5191-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.514655 4824 generic.go:334] "Generic (PLEG): container finished" podID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerID="bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c" exitCode=0 Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.515150 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ztmfn" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.515391 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ztmfn" event={"ID":"bfb91335-3f3a-45da-86c0-4decb3bd5191","Type":"ContainerDied","Data":"bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c"} Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.515432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ztmfn" event={"ID":"bfb91335-3f3a-45da-86c0-4decb3bd5191","Type":"ContainerDied","Data":"368e1d546ae8ef4fa4478c88c223927169ad285243efb5c397351f9f278cbaff"} Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.515450 4824 scope.go:117] "RemoveContainer" containerID="bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.538092 4824 scope.go:117] "RemoveContainer" containerID="571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.561125 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ztmfn"] Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.575919 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ztmfn"] Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.582830 4824 scope.go:117] "RemoveContainer" containerID="be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.617492 4824 scope.go:117] "RemoveContainer" containerID="bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c" Nov 24 14:03:53 crc kubenswrapper[4824]: E1124 14:03:53.625220 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c\": container with ID starting with bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c not found: ID does not exist" containerID="bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.625283 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c"} err="failed to get container status \"bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c\": rpc error: code = NotFound desc = could not find container \"bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c\": container with ID starting with bc511bd3dd291c155d1d410a0f5a414db9b91605429547297f5fefa5bdd1242c not found: ID does not exist" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.625319 4824 scope.go:117] "RemoveContainer" containerID="571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b" Nov 24 14:03:53 crc kubenswrapper[4824]: E1124 14:03:53.625940 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b\": container with ID starting with 571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b not found: ID does not exist" containerID="571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.625966 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b"} err="failed to get container status \"571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b\": rpc error: code = NotFound desc = could not find container \"571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b\": container with ID starting with 571e6ad24301227d0637f45711ff1f6d0ca1f79108d86f555628cc8258349f2b not found: ID does not exist" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.625982 4824 scope.go:117] "RemoveContainer" containerID="be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc" Nov 24 14:03:53 crc kubenswrapper[4824]: E1124 14:03:53.626471 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc\": container with ID starting with be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc not found: ID does not exist" containerID="be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc" Nov 24 14:03:53 crc kubenswrapper[4824]: I1124 14:03:53.626522 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc"} err="failed to get container status \"be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc\": rpc error: code = NotFound desc = could not find container \"be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc\": container with ID starting with be0955fc1a9741faee3073f079718ce67188335d0713cfe917748d9e80dcf2fc not found: ID does not exist" Nov 24 14:03:54 crc kubenswrapper[4824]: I1124 14:03:54.531963 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h8s6m" podUID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerName="registry-server" containerID="cri-o://4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de" gracePeriod=2 Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.021016 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfb91335-3f3a-45da-86c0-4decb3bd5191" path="/var/lib/kubelet/pods/bfb91335-3f3a-45da-86c0-4decb3bd5191/volumes" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.043730 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.203244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-catalog-content\") pod \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.203323 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-utilities\") pod \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.203388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w568b\" (UniqueName: \"kubernetes.io/projected/54e5f8c6-26f1-4318-b507-d71defd8ee8f-kube-api-access-w568b\") pod \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\" (UID: \"54e5f8c6-26f1-4318-b507-d71defd8ee8f\") " Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.204460 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-utilities" (OuterVolumeSpecName: "utilities") pod "54e5f8c6-26f1-4318-b507-d71defd8ee8f" (UID: "54e5f8c6-26f1-4318-b507-d71defd8ee8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.209980 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e5f8c6-26f1-4318-b507-d71defd8ee8f-kube-api-access-w568b" (OuterVolumeSpecName: "kube-api-access-w568b") pod "54e5f8c6-26f1-4318-b507-d71defd8ee8f" (UID: "54e5f8c6-26f1-4318-b507-d71defd8ee8f"). InnerVolumeSpecName "kube-api-access-w568b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.221613 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54e5f8c6-26f1-4318-b507-d71defd8ee8f" (UID: "54e5f8c6-26f1-4318-b507-d71defd8ee8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.308746 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.309084 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e5f8c6-26f1-4318-b507-d71defd8ee8f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.309119 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w568b\" (UniqueName: \"kubernetes.io/projected/54e5f8c6-26f1-4318-b507-d71defd8ee8f-kube-api-access-w568b\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.542773 4824 generic.go:334] "Generic (PLEG): container finished" podID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerID="4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de" exitCode=0 Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.542842 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h8s6m" event={"ID":"54e5f8c6-26f1-4318-b507-d71defd8ee8f","Type":"ContainerDied","Data":"4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de"} Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.542868 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h8s6m" event={"ID":"54e5f8c6-26f1-4318-b507-d71defd8ee8f","Type":"ContainerDied","Data":"c619578116dd74f567a1f24b7a0fd8639ca89cf7f809f343bb6e52ae31296f1c"} Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.542906 4824 scope.go:117] "RemoveContainer" containerID="4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.542898 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h8s6m" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.581517 4824 scope.go:117] "RemoveContainer" containerID="390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.596243 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h8s6m"] Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.623519 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h8s6m"] Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.626223 4824 scope.go:117] "RemoveContainer" containerID="7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.693675 4824 scope.go:117] "RemoveContainer" containerID="4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de" Nov 24 14:03:55 crc kubenswrapper[4824]: E1124 14:03:55.696223 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de\": container with ID starting with 4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de not found: ID does not exist" containerID="4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.696569 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de"} err="failed to get container status \"4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de\": rpc error: code = NotFound desc = could not find container \"4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de\": container with ID starting with 4eee7cf0a125e760adffac01c48877725026ad5befd676bc4784645661fbd6de not found: ID does not exist" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.696603 4824 scope.go:117] "RemoveContainer" containerID="390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9" Nov 24 14:03:55 crc kubenswrapper[4824]: E1124 14:03:55.696976 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9\": container with ID starting with 390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9 not found: ID does not exist" containerID="390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.697011 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9"} err="failed to get container status \"390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9\": rpc error: code = NotFound desc = could not find container \"390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9\": container with ID starting with 390dfa2c6e7eb3f1b4934ca4e50360e1f0d8fe36161aa4f781858aa1405c4ce9 not found: ID does not exist" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.697034 4824 scope.go:117] "RemoveContainer" containerID="7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6" Nov 24 14:03:55 crc kubenswrapper[4824]: E1124 14:03:55.697381 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6\": container with ID starting with 7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6 not found: ID does not exist" containerID="7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6" Nov 24 14:03:55 crc kubenswrapper[4824]: I1124 14:03:55.697425 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6"} err="failed to get container status \"7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6\": rpc error: code = NotFound desc = could not find container \"7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6\": container with ID starting with 7ab7be0ecbae88f1ef6786ea946e70d071c0478b9f565072a5b9eda8d56781c6 not found: ID does not exist" Nov 24 14:03:57 crc kubenswrapper[4824]: I1124 14:03:57.020794 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" path="/var/lib/kubelet/pods/54e5f8c6-26f1-4318-b507-d71defd8ee8f/volumes" Nov 24 14:03:59 crc kubenswrapper[4824]: I1124 14:03:59.548483 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:03:59 crc kubenswrapper[4824]: I1124 14:03:59.603166 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:04:00 crc kubenswrapper[4824]: I1124 14:04:00.610656 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tc5fb"] Nov 24 14:04:00 crc kubenswrapper[4824]: I1124 14:04:00.610958 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tc5fb" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="registry-server" containerID="cri-o://2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290" gracePeriod=2 Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.062778 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.229116 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-catalog-content\") pod \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.229206 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hm7w\" (UniqueName: \"kubernetes.io/projected/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-kube-api-access-4hm7w\") pod \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.229327 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-utilities\") pod \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\" (UID: \"af996432-3c6f-4aaa-a216-2e1bc56d8bc0\") " Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.230156 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-utilities" (OuterVolumeSpecName: "utilities") pod "af996432-3c6f-4aaa-a216-2e1bc56d8bc0" (UID: "af996432-3c6f-4aaa-a216-2e1bc56d8bc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.236686 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-kube-api-access-4hm7w" (OuterVolumeSpecName: "kube-api-access-4hm7w") pod "af996432-3c6f-4aaa-a216-2e1bc56d8bc0" (UID: "af996432-3c6f-4aaa-a216-2e1bc56d8bc0"). InnerVolumeSpecName "kube-api-access-4hm7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.300949 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af996432-3c6f-4aaa-a216-2e1bc56d8bc0" (UID: "af996432-3c6f-4aaa-a216-2e1bc56d8bc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.332009 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hm7w\" (UniqueName: \"kubernetes.io/projected/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-kube-api-access-4hm7w\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.332040 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.332053 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af996432-3c6f-4aaa-a216-2e1bc56d8bc0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.607450 4824 generic.go:334] "Generic (PLEG): container finished" podID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerID="2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290" exitCode=0 Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.607491 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tc5fb" event={"ID":"af996432-3c6f-4aaa-a216-2e1bc56d8bc0","Type":"ContainerDied","Data":"2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290"} Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.607577 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tc5fb" event={"ID":"af996432-3c6f-4aaa-a216-2e1bc56d8bc0","Type":"ContainerDied","Data":"74efa369006a17168703574eec6547463f432f6587dafce581201d461f0af3c7"} Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.607602 4824 scope.go:117] "RemoveContainer" containerID="2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.608204 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tc5fb" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.643363 4824 scope.go:117] "RemoveContainer" containerID="2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.672874 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tc5fb"] Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.691229 4824 scope.go:117] "RemoveContainer" containerID="7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.692133 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tc5fb"] Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.741897 4824 scope.go:117] "RemoveContainer" containerID="2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290" Nov 24 14:04:01 crc kubenswrapper[4824]: E1124 14:04:01.742672 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290\": container with ID starting with 2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290 not found: ID does not exist" containerID="2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.742790 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290"} err="failed to get container status \"2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290\": rpc error: code = NotFound desc = could not find container \"2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290\": container with ID starting with 2e01e3b2440931d8f706b47b0dbf1649ce0d0cbe3d60b058d4ef442bf4828290 not found: ID does not exist" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.742916 4824 scope.go:117] "RemoveContainer" containerID="2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f" Nov 24 14:04:01 crc kubenswrapper[4824]: E1124 14:04:01.743543 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f\": container with ID starting with 2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f not found: ID does not exist" containerID="2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.743623 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f"} err="failed to get container status \"2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f\": rpc error: code = NotFound desc = could not find container \"2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f\": container with ID starting with 2d3e373644a33bd8dde70ea947f8b9bf035c5547d4e21dee827f0d96c58d289f not found: ID does not exist" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.743658 4824 scope.go:117] "RemoveContainer" containerID="7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17" Nov 24 14:04:01 crc kubenswrapper[4824]: E1124 14:04:01.744035 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17\": container with ID starting with 7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17 not found: ID does not exist" containerID="7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17" Nov 24 14:04:01 crc kubenswrapper[4824]: I1124 14:04:01.744068 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17"} err="failed to get container status \"7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17\": rpc error: code = NotFound desc = could not find container \"7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17\": container with ID starting with 7c8c707b8a5b3a9192900a33e3fcee74822a3374b7faa0f07fc8bbf31d0c3c17 not found: ID does not exist" Nov 24 14:04:03 crc kubenswrapper[4824]: I1124 14:04:03.023737 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" path="/var/lib/kubelet/pods/af996432-3c6f-4aaa-a216-2e1bc56d8bc0/volumes" Nov 24 14:05:40 crc kubenswrapper[4824]: I1124 14:05:40.788269 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:05:40 crc kubenswrapper[4824]: I1124 14:05:40.789213 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:06:10 crc kubenswrapper[4824]: I1124 14:06:10.787474 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:06:10 crc kubenswrapper[4824]: I1124 14:06:10.788073 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:06:40 crc kubenswrapper[4824]: I1124 14:06:40.787385 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:06:40 crc kubenswrapper[4824]: I1124 14:06:40.787997 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:06:40 crc kubenswrapper[4824]: I1124 14:06:40.788044 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 14:06:40 crc kubenswrapper[4824]: I1124 14:06:40.788838 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:06:40 crc kubenswrapper[4824]: I1124 14:06:40.788893 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" gracePeriod=600 Nov 24 14:06:40 crc kubenswrapper[4824]: E1124 14:06:40.913182 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:06:41 crc kubenswrapper[4824]: I1124 14:06:41.127088 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" exitCode=0 Nov 24 14:06:41 crc kubenswrapper[4824]: I1124 14:06:41.127134 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c"} Nov 24 14:06:41 crc kubenswrapper[4824]: I1124 14:06:41.127187 4824 scope.go:117] "RemoveContainer" containerID="9a8bedcb7448d68b9ae3979be0e7a2d503d2d1ac96679328d965b1f472005e8d" Nov 24 14:06:41 crc kubenswrapper[4824]: I1124 14:06:41.128128 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:06:41 crc kubenswrapper[4824]: E1124 14:06:41.128540 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:06:55 crc kubenswrapper[4824]: I1124 14:06:55.011021 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:06:55 crc kubenswrapper[4824]: E1124 14:06:55.011871 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:07:08 crc kubenswrapper[4824]: I1124 14:07:08.010580 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:07:08 crc kubenswrapper[4824]: E1124 14:07:08.011189 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:07:19 crc kubenswrapper[4824]: I1124 14:07:19.030344 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:07:19 crc kubenswrapper[4824]: E1124 14:07:19.032543 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:07:30 crc kubenswrapper[4824]: I1124 14:07:30.010161 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:07:30 crc kubenswrapper[4824]: E1124 14:07:30.010853 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:07:45 crc kubenswrapper[4824]: I1124 14:07:45.011307 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:07:45 crc kubenswrapper[4824]: E1124 14:07:45.012244 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:07:59 crc kubenswrapper[4824]: I1124 14:07:59.019862 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:07:59 crc kubenswrapper[4824]: E1124 14:07:59.020645 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:08:13 crc kubenswrapper[4824]: I1124 14:08:13.010744 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:08:13 crc kubenswrapper[4824]: E1124 14:08:13.011576 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:08:24 crc kubenswrapper[4824]: I1124 14:08:24.010918 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:08:24 crc kubenswrapper[4824]: E1124 14:08:24.011959 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:08:38 crc kubenswrapper[4824]: I1124 14:08:38.010218 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:08:38 crc kubenswrapper[4824]: E1124 14:08:38.011225 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:08:49 crc kubenswrapper[4824]: I1124 14:08:49.016683 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:08:49 crc kubenswrapper[4824]: E1124 14:08:49.017539 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.695223 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qjxkc"] Nov 24 14:08:50 crc kubenswrapper[4824]: E1124 14:08:50.695994 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerName="extract-utilities" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696011 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerName="extract-utilities" Nov 24 14:08:50 crc kubenswrapper[4824]: E1124 14:08:50.696038 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="extract-content" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696044 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="extract-content" Nov 24 14:08:50 crc kubenswrapper[4824]: E1124 14:08:50.696053 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="extract-utilities" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696059 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="extract-utilities" Nov 24 14:08:50 crc kubenswrapper[4824]: E1124 14:08:50.696073 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerName="extract-utilities" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696079 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerName="extract-utilities" Nov 24 14:08:50 crc kubenswrapper[4824]: E1124 14:08:50.696093 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerName="registry-server" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696100 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerName="registry-server" Nov 24 14:08:50 crc kubenswrapper[4824]: E1124 14:08:50.696118 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="registry-server" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696125 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="registry-server" Nov 24 14:08:50 crc kubenswrapper[4824]: E1124 14:08:50.696134 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerName="extract-content" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696141 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerName="extract-content" Nov 24 14:08:50 crc kubenswrapper[4824]: E1124 14:08:50.696153 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerName="registry-server" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696160 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerName="registry-server" Nov 24 14:08:50 crc kubenswrapper[4824]: E1124 14:08:50.696181 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerName="extract-content" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696190 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerName="extract-content" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696410 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb91335-3f3a-45da-86c0-4decb3bd5191" containerName="registry-server" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696427 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="54e5f8c6-26f1-4318-b507-d71defd8ee8f" containerName="registry-server" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.696450 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="af996432-3c6f-4aaa-a216-2e1bc56d8bc0" containerName="registry-server" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.697893 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.724441 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qjxkc"] Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.809823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-catalog-content\") pod \"redhat-operators-qjxkc\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.810226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x5lh\" (UniqueName: \"kubernetes.io/projected/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-kube-api-access-7x5lh\") pod \"redhat-operators-qjxkc\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.810330 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-utilities\") pod \"redhat-operators-qjxkc\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.911478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x5lh\" (UniqueName: \"kubernetes.io/projected/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-kube-api-access-7x5lh\") pod \"redhat-operators-qjxkc\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.911575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-utilities\") pod \"redhat-operators-qjxkc\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.911622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-catalog-content\") pod \"redhat-operators-qjxkc\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.912141 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-catalog-content\") pod \"redhat-operators-qjxkc\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.912144 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-utilities\") pod \"redhat-operators-qjxkc\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:50 crc kubenswrapper[4824]: I1124 14:08:50.974828 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x5lh\" (UniqueName: \"kubernetes.io/projected/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-kube-api-access-7x5lh\") pod \"redhat-operators-qjxkc\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:51 crc kubenswrapper[4824]: I1124 14:08:51.047949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:08:52 crc kubenswrapper[4824]: I1124 14:08:52.491118 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qjxkc"] Nov 24 14:08:53 crc kubenswrapper[4824]: I1124 14:08:53.331747 4824 generic.go:334] "Generic (PLEG): container finished" podID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerID="c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181" exitCode=0 Nov 24 14:08:53 crc kubenswrapper[4824]: I1124 14:08:53.331850 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjxkc" event={"ID":"e12f5d8e-d074-4e91-bc78-9b1db0a86acb","Type":"ContainerDied","Data":"c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181"} Nov 24 14:08:53 crc kubenswrapper[4824]: I1124 14:08:53.332114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjxkc" event={"ID":"e12f5d8e-d074-4e91-bc78-9b1db0a86acb","Type":"ContainerStarted","Data":"de3ab25bacbfeabb36d6246d0fcca577e7560db62e14d349953f92164c4fe96f"} Nov 24 14:08:53 crc kubenswrapper[4824]: I1124 14:08:53.334366 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:08:55 crc kubenswrapper[4824]: I1124 14:08:55.352990 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjxkc" event={"ID":"e12f5d8e-d074-4e91-bc78-9b1db0a86acb","Type":"ContainerStarted","Data":"9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197"} Nov 24 14:09:00 crc kubenswrapper[4824]: I1124 14:09:00.410113 4824 generic.go:334] "Generic (PLEG): container finished" podID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerID="9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197" exitCode=0 Nov 24 14:09:00 crc kubenswrapper[4824]: I1124 14:09:00.410289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjxkc" event={"ID":"e12f5d8e-d074-4e91-bc78-9b1db0a86acb","Type":"ContainerDied","Data":"9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197"} Nov 24 14:09:01 crc kubenswrapper[4824]: I1124 14:09:01.420245 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjxkc" event={"ID":"e12f5d8e-d074-4e91-bc78-9b1db0a86acb","Type":"ContainerStarted","Data":"3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3"} Nov 24 14:09:01 crc kubenswrapper[4824]: I1124 14:09:01.441342 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qjxkc" podStartSLOduration=3.948756357 podStartE2EDuration="11.441310532s" podCreationTimestamp="2025-11-24 14:08:50 +0000 UTC" firstStartedPulling="2025-11-24 14:08:53.334127313 +0000 UTC m=+3274.973666623" lastFinishedPulling="2025-11-24 14:09:00.826681488 +0000 UTC m=+3282.466220798" observedRunningTime="2025-11-24 14:09:01.436287046 +0000 UTC m=+3283.075826366" watchObservedRunningTime="2025-11-24 14:09:01.441310532 +0000 UTC m=+3283.080849852" Nov 24 14:09:03 crc kubenswrapper[4824]: I1124 14:09:03.010156 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:09:03 crc kubenswrapper[4824]: E1124 14:09:03.010747 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:09:11 crc kubenswrapper[4824]: I1124 14:09:11.048362 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:09:11 crc kubenswrapper[4824]: I1124 14:09:11.048770 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:09:11 crc kubenswrapper[4824]: I1124 14:09:11.110204 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:09:11 crc kubenswrapper[4824]: I1124 14:09:11.547688 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:09:11 crc kubenswrapper[4824]: I1124 14:09:11.594727 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qjxkc"] Nov 24 14:09:13 crc kubenswrapper[4824]: I1124 14:09:13.518470 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qjxkc" podUID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerName="registry-server" containerID="cri-o://3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3" gracePeriod=2 Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.224993 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.361423 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x5lh\" (UniqueName: \"kubernetes.io/projected/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-kube-api-access-7x5lh\") pod \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.361651 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-catalog-content\") pod \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.361697 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-utilities\") pod \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\" (UID: \"e12f5d8e-d074-4e91-bc78-9b1db0a86acb\") " Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.362756 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-utilities" (OuterVolumeSpecName: "utilities") pod "e12f5d8e-d074-4e91-bc78-9b1db0a86acb" (UID: "e12f5d8e-d074-4e91-bc78-9b1db0a86acb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.380061 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-kube-api-access-7x5lh" (OuterVolumeSpecName: "kube-api-access-7x5lh") pod "e12f5d8e-d074-4e91-bc78-9b1db0a86acb" (UID: "e12f5d8e-d074-4e91-bc78-9b1db0a86acb"). InnerVolumeSpecName "kube-api-access-7x5lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.443127 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e12f5d8e-d074-4e91-bc78-9b1db0a86acb" (UID: "e12f5d8e-d074-4e91-bc78-9b1db0a86acb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.463763 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.464055 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.464118 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x5lh\" (UniqueName: \"kubernetes.io/projected/e12f5d8e-d074-4e91-bc78-9b1db0a86acb-kube-api-access-7x5lh\") on node \"crc\" DevicePath \"\"" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.529235 4824 generic.go:334] "Generic (PLEG): container finished" podID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerID="3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3" exitCode=0 Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.529282 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjxkc" event={"ID":"e12f5d8e-d074-4e91-bc78-9b1db0a86acb","Type":"ContainerDied","Data":"3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3"} Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.529314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjxkc" event={"ID":"e12f5d8e-d074-4e91-bc78-9b1db0a86acb","Type":"ContainerDied","Data":"de3ab25bacbfeabb36d6246d0fcca577e7560db62e14d349953f92164c4fe96f"} Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.529337 4824 scope.go:117] "RemoveContainer" containerID="3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.531625 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjxkc" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.569834 4824 scope.go:117] "RemoveContainer" containerID="9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.580434 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qjxkc"] Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.594343 4824 scope.go:117] "RemoveContainer" containerID="c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.599296 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qjxkc"] Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.645196 4824 scope.go:117] "RemoveContainer" containerID="3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3" Nov 24 14:09:14 crc kubenswrapper[4824]: E1124 14:09:14.645567 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3\": container with ID starting with 3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3 not found: ID does not exist" containerID="3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.645594 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3"} err="failed to get container status \"3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3\": rpc error: code = NotFound desc = could not find container \"3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3\": container with ID starting with 3347ba714547cc0a8a233cebdbfb6715af318f961fc5ad3b7da5319887e005e3 not found: ID does not exist" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.645616 4824 scope.go:117] "RemoveContainer" containerID="9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197" Nov 24 14:09:14 crc kubenswrapper[4824]: E1124 14:09:14.645978 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197\": container with ID starting with 9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197 not found: ID does not exist" containerID="9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.646007 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197"} err="failed to get container status \"9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197\": rpc error: code = NotFound desc = could not find container \"9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197\": container with ID starting with 9611bad0a084a260e1190a5020ddf28e773338ed824f4993c17b6d346940f197 not found: ID does not exist" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.646019 4824 scope.go:117] "RemoveContainer" containerID="c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181" Nov 24 14:09:14 crc kubenswrapper[4824]: E1124 14:09:14.646214 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181\": container with ID starting with c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181 not found: ID does not exist" containerID="c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181" Nov 24 14:09:14 crc kubenswrapper[4824]: I1124 14:09:14.646233 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181"} err="failed to get container status \"c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181\": rpc error: code = NotFound desc = could not find container \"c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181\": container with ID starting with c6f721d3c2a2f50f06f62331c4d5180bcb5f8dcf75ee43c3f887b437208c2181 not found: ID does not exist" Nov 24 14:09:15 crc kubenswrapper[4824]: I1124 14:09:15.021283 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" path="/var/lib/kubelet/pods/e12f5d8e-d074-4e91-bc78-9b1db0a86acb/volumes" Nov 24 14:09:16 crc kubenswrapper[4824]: I1124 14:09:16.010109 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:09:16 crc kubenswrapper[4824]: E1124 14:09:16.010618 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:09:29 crc kubenswrapper[4824]: I1124 14:09:29.025908 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:09:29 crc kubenswrapper[4824]: E1124 14:09:29.026620 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:09:42 crc kubenswrapper[4824]: I1124 14:09:42.010585 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:09:42 crc kubenswrapper[4824]: E1124 14:09:42.011354 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:09:57 crc kubenswrapper[4824]: I1124 14:09:57.013450 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:09:57 crc kubenswrapper[4824]: E1124 14:09:57.015445 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:10:09 crc kubenswrapper[4824]: I1124 14:10:09.024072 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:10:09 crc kubenswrapper[4824]: E1124 14:10:09.024993 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:10:22 crc kubenswrapper[4824]: I1124 14:10:22.010096 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:10:22 crc kubenswrapper[4824]: E1124 14:10:22.010719 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:10:36 crc kubenswrapper[4824]: I1124 14:10:36.009580 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:10:36 crc kubenswrapper[4824]: E1124 14:10:36.010379 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:10:47 crc kubenswrapper[4824]: I1124 14:10:47.010357 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:10:47 crc kubenswrapper[4824]: E1124 14:10:47.011139 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:11:00 crc kubenswrapper[4824]: I1124 14:11:00.009969 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:11:00 crc kubenswrapper[4824]: E1124 14:11:00.010711 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:11:13 crc kubenswrapper[4824]: I1124 14:11:13.010159 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:11:13 crc kubenswrapper[4824]: E1124 14:11:13.011038 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:11:24 crc kubenswrapper[4824]: I1124 14:11:24.012065 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:11:24 crc kubenswrapper[4824]: E1124 14:11:24.012915 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:11:38 crc kubenswrapper[4824]: I1124 14:11:38.010401 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:11:38 crc kubenswrapper[4824]: E1124 14:11:38.011215 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:11:52 crc kubenswrapper[4824]: I1124 14:11:52.010658 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:11:53 crc kubenswrapper[4824]: I1124 14:11:53.093364 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"4e0b99c526a62db8017cf1ab4e1d16ce7faea62b4707845d45ae09b4bd27370f"} Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.038859 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ztll7"] Nov 24 14:13:28 crc kubenswrapper[4824]: E1124 14:13:28.040991 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerName="registry-server" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.041110 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerName="registry-server" Nov 24 14:13:28 crc kubenswrapper[4824]: E1124 14:13:28.041243 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerName="extract-utilities" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.041353 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerName="extract-utilities" Nov 24 14:13:28 crc kubenswrapper[4824]: E1124 14:13:28.041448 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerName="extract-content" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.041532 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerName="extract-content" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.041877 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e12f5d8e-d074-4e91-bc78-9b1db0a86acb" containerName="registry-server" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.048460 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.056423 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztll7"] Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.112439 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dbqq\" (UniqueName: \"kubernetes.io/projected/bdab8cda-1502-4cbe-945c-8f34818d8190-kube-api-access-5dbqq\") pod \"redhat-marketplace-ztll7\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.112498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-utilities\") pod \"redhat-marketplace-ztll7\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.112621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-catalog-content\") pod \"redhat-marketplace-ztll7\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.213404 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-catalog-content\") pod \"redhat-marketplace-ztll7\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.213541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dbqq\" (UniqueName: \"kubernetes.io/projected/bdab8cda-1502-4cbe-945c-8f34818d8190-kube-api-access-5dbqq\") pod \"redhat-marketplace-ztll7\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.213570 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-utilities\") pod \"redhat-marketplace-ztll7\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.213888 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-catalog-content\") pod \"redhat-marketplace-ztll7\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.214239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-utilities\") pod \"redhat-marketplace-ztll7\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.242072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dbqq\" (UniqueName: \"kubernetes.io/projected/bdab8cda-1502-4cbe-945c-8f34818d8190-kube-api-access-5dbqq\") pod \"redhat-marketplace-ztll7\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.386776 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.943261 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztll7"] Nov 24 14:13:28 crc kubenswrapper[4824]: I1124 14:13:28.974042 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztll7" event={"ID":"bdab8cda-1502-4cbe-945c-8f34818d8190","Type":"ContainerStarted","Data":"b8bb41eeaaa4b32e03b18d5ce62bc7c178ef29c6523999ffa759bbac3c562b3e"} Nov 24 14:13:29 crc kubenswrapper[4824]: I1124 14:13:29.985534 4824 generic.go:334] "Generic (PLEG): container finished" podID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerID="1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e" exitCode=0 Nov 24 14:13:29 crc kubenswrapper[4824]: I1124 14:13:29.986181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztll7" event={"ID":"bdab8cda-1502-4cbe-945c-8f34818d8190","Type":"ContainerDied","Data":"1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e"} Nov 24 14:13:30 crc kubenswrapper[4824]: I1124 14:13:30.996734 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztll7" event={"ID":"bdab8cda-1502-4cbe-945c-8f34818d8190","Type":"ContainerStarted","Data":"8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5"} Nov 24 14:13:33 crc kubenswrapper[4824]: I1124 14:13:33.015395 4824 generic.go:334] "Generic (PLEG): container finished" podID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerID="8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5" exitCode=0 Nov 24 14:13:33 crc kubenswrapper[4824]: I1124 14:13:33.020564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztll7" event={"ID":"bdab8cda-1502-4cbe-945c-8f34818d8190","Type":"ContainerDied","Data":"8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5"} Nov 24 14:13:34 crc kubenswrapper[4824]: I1124 14:13:34.027541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztll7" event={"ID":"bdab8cda-1502-4cbe-945c-8f34818d8190","Type":"ContainerStarted","Data":"dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef"} Nov 24 14:13:34 crc kubenswrapper[4824]: I1124 14:13:34.059900 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ztll7" podStartSLOduration=2.616704416 podStartE2EDuration="6.059883468s" podCreationTimestamp="2025-11-24 14:13:28 +0000 UTC" firstStartedPulling="2025-11-24 14:13:29.988161869 +0000 UTC m=+3551.627701179" lastFinishedPulling="2025-11-24 14:13:33.431340921 +0000 UTC m=+3555.070880231" observedRunningTime="2025-11-24 14:13:34.054819234 +0000 UTC m=+3555.694358554" watchObservedRunningTime="2025-11-24 14:13:34.059883468 +0000 UTC m=+3555.699422778" Nov 24 14:13:38 crc kubenswrapper[4824]: I1124 14:13:38.387324 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:38 crc kubenswrapper[4824]: I1124 14:13:38.387943 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:38 crc kubenswrapper[4824]: I1124 14:13:38.445006 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:39 crc kubenswrapper[4824]: I1124 14:13:39.120201 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:39 crc kubenswrapper[4824]: I1124 14:13:39.177288 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztll7"] Nov 24 14:13:41 crc kubenswrapper[4824]: I1124 14:13:41.088720 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ztll7" podUID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerName="registry-server" containerID="cri-o://dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef" gracePeriod=2 Nov 24 14:13:41 crc kubenswrapper[4824]: I1124 14:13:41.886302 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:41 crc kubenswrapper[4824]: I1124 14:13:41.990989 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-utilities\") pod \"bdab8cda-1502-4cbe-945c-8f34818d8190\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " Nov 24 14:13:41 crc kubenswrapper[4824]: I1124 14:13:41.991526 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-catalog-content\") pod \"bdab8cda-1502-4cbe-945c-8f34818d8190\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " Nov 24 14:13:41 crc kubenswrapper[4824]: I1124 14:13:41.991751 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dbqq\" (UniqueName: \"kubernetes.io/projected/bdab8cda-1502-4cbe-945c-8f34818d8190-kube-api-access-5dbqq\") pod \"bdab8cda-1502-4cbe-945c-8f34818d8190\" (UID: \"bdab8cda-1502-4cbe-945c-8f34818d8190\") " Nov 24 14:13:41 crc kubenswrapper[4824]: I1124 14:13:41.994039 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-utilities" (OuterVolumeSpecName: "utilities") pod "bdab8cda-1502-4cbe-945c-8f34818d8190" (UID: "bdab8cda-1502-4cbe-945c-8f34818d8190"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.001982 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdab8cda-1502-4cbe-945c-8f34818d8190-kube-api-access-5dbqq" (OuterVolumeSpecName: "kube-api-access-5dbqq") pod "bdab8cda-1502-4cbe-945c-8f34818d8190" (UID: "bdab8cda-1502-4cbe-945c-8f34818d8190"). InnerVolumeSpecName "kube-api-access-5dbqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.011434 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdab8cda-1502-4cbe-945c-8f34818d8190" (UID: "bdab8cda-1502-4cbe-945c-8f34818d8190"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.094085 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dbqq\" (UniqueName: \"kubernetes.io/projected/bdab8cda-1502-4cbe-945c-8f34818d8190-kube-api-access-5dbqq\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.094118 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.094128 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdab8cda-1502-4cbe-945c-8f34818d8190-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.099903 4824 generic.go:334] "Generic (PLEG): container finished" podID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerID="dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef" exitCode=0 Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.099943 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztll7" event={"ID":"bdab8cda-1502-4cbe-945c-8f34818d8190","Type":"ContainerDied","Data":"dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef"} Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.099969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztll7" event={"ID":"bdab8cda-1502-4cbe-945c-8f34818d8190","Type":"ContainerDied","Data":"b8bb41eeaaa4b32e03b18d5ce62bc7c178ef29c6523999ffa759bbac3c562b3e"} Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.099986 4824 scope.go:117] "RemoveContainer" containerID="dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.100347 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztll7" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.126777 4824 scope.go:117] "RemoveContainer" containerID="8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.145216 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztll7"] Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.155367 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztll7"] Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.189846 4824 scope.go:117] "RemoveContainer" containerID="1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.227925 4824 scope.go:117] "RemoveContainer" containerID="dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef" Nov 24 14:13:42 crc kubenswrapper[4824]: E1124 14:13:42.230935 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef\": container with ID starting with dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef not found: ID does not exist" containerID="dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.230979 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef"} err="failed to get container status \"dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef\": rpc error: code = NotFound desc = could not find container \"dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef\": container with ID starting with dd0a92e4192ac3df79e212a117803257b2f6d32864a50f75d5575a39c1f79aef not found: ID does not exist" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.231008 4824 scope.go:117] "RemoveContainer" containerID="8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5" Nov 24 14:13:42 crc kubenswrapper[4824]: E1124 14:13:42.231417 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5\": container with ID starting with 8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5 not found: ID does not exist" containerID="8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.231450 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5"} err="failed to get container status \"8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5\": rpc error: code = NotFound desc = could not find container \"8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5\": container with ID starting with 8b4d25bbe5e81ba5612ade0cc54d40f1dbd529627db1a0953ed8a426e3b15bf5 not found: ID does not exist" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.231489 4824 scope.go:117] "RemoveContainer" containerID="1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e" Nov 24 14:13:42 crc kubenswrapper[4824]: E1124 14:13:42.231682 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e\": container with ID starting with 1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e not found: ID does not exist" containerID="1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e" Nov 24 14:13:42 crc kubenswrapper[4824]: I1124 14:13:42.231717 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e"} err="failed to get container status \"1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e\": rpc error: code = NotFound desc = could not find container \"1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e\": container with ID starting with 1a34c5581f619f54aa312a5a1a49f1075ea1bdc81cbc98dd681510479c43119e not found: ID does not exist" Nov 24 14:13:43 crc kubenswrapper[4824]: I1124 14:13:43.021525 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdab8cda-1502-4cbe-945c-8f34818d8190" path="/var/lib/kubelet/pods/bdab8cda-1502-4cbe-945c-8f34818d8190/volumes" Nov 24 14:14:10 crc kubenswrapper[4824]: I1124 14:14:10.789044 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:14:10 crc kubenswrapper[4824]: I1124 14:14:10.789681 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.313821 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nh4bn"] Nov 24 14:14:35 crc kubenswrapper[4824]: E1124 14:14:35.314544 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerName="extract-utilities" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.314556 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerName="extract-utilities" Nov 24 14:14:35 crc kubenswrapper[4824]: E1124 14:14:35.314603 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerName="extract-content" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.314609 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerName="extract-content" Nov 24 14:14:35 crc kubenswrapper[4824]: E1124 14:14:35.314624 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerName="registry-server" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.314631 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerName="registry-server" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.314801 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdab8cda-1502-4cbe-945c-8f34818d8190" containerName="registry-server" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.316073 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.336546 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nh4bn"] Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.374308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe27b019-f6cf-4eec-83df-ef23b6483cf8-utilities\") pod \"community-operators-nh4bn\" (UID: \"fe27b019-f6cf-4eec-83df-ef23b6483cf8\") " pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.374575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe27b019-f6cf-4eec-83df-ef23b6483cf8-catalog-content\") pod \"community-operators-nh4bn\" (UID: \"fe27b019-f6cf-4eec-83df-ef23b6483cf8\") " pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.374746 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt9rx\" (UniqueName: \"kubernetes.io/projected/fe27b019-f6cf-4eec-83df-ef23b6483cf8-kube-api-access-bt9rx\") pod \"community-operators-nh4bn\" (UID: \"fe27b019-f6cf-4eec-83df-ef23b6483cf8\") " pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.476888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe27b019-f6cf-4eec-83df-ef23b6483cf8-utilities\") pod \"community-operators-nh4bn\" (UID: \"fe27b019-f6cf-4eec-83df-ef23b6483cf8\") " pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.476988 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe27b019-f6cf-4eec-83df-ef23b6483cf8-catalog-content\") pod \"community-operators-nh4bn\" (UID: \"fe27b019-f6cf-4eec-83df-ef23b6483cf8\") " pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.477042 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt9rx\" (UniqueName: \"kubernetes.io/projected/fe27b019-f6cf-4eec-83df-ef23b6483cf8-kube-api-access-bt9rx\") pod \"community-operators-nh4bn\" (UID: \"fe27b019-f6cf-4eec-83df-ef23b6483cf8\") " pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.477655 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe27b019-f6cf-4eec-83df-ef23b6483cf8-utilities\") pod \"community-operators-nh4bn\" (UID: \"fe27b019-f6cf-4eec-83df-ef23b6483cf8\") " pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.477694 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe27b019-f6cf-4eec-83df-ef23b6483cf8-catalog-content\") pod \"community-operators-nh4bn\" (UID: \"fe27b019-f6cf-4eec-83df-ef23b6483cf8\") " pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.498622 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt9rx\" (UniqueName: \"kubernetes.io/projected/fe27b019-f6cf-4eec-83df-ef23b6483cf8-kube-api-access-bt9rx\") pod \"community-operators-nh4bn\" (UID: \"fe27b019-f6cf-4eec-83df-ef23b6483cf8\") " pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:35 crc kubenswrapper[4824]: I1124 14:14:35.646045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:36 crc kubenswrapper[4824]: I1124 14:14:36.282354 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nh4bn"] Nov 24 14:14:36 crc kubenswrapper[4824]: I1124 14:14:36.570921 4824 generic.go:334] "Generic (PLEG): container finished" podID="fe27b019-f6cf-4eec-83df-ef23b6483cf8" containerID="276a7bb713289727a043eb3a4279039739d9b4a2c40f3feea284724fe02a29f7" exitCode=0 Nov 24 14:14:36 crc kubenswrapper[4824]: I1124 14:14:36.571011 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nh4bn" event={"ID":"fe27b019-f6cf-4eec-83df-ef23b6483cf8","Type":"ContainerDied","Data":"276a7bb713289727a043eb3a4279039739d9b4a2c40f3feea284724fe02a29f7"} Nov 24 14:14:36 crc kubenswrapper[4824]: I1124 14:14:36.572152 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nh4bn" event={"ID":"fe27b019-f6cf-4eec-83df-ef23b6483cf8","Type":"ContainerStarted","Data":"05ae382fc3cf30ec928ef5aec710f5f867e7bc12d071841f73127247573a4715"} Nov 24 14:14:36 crc kubenswrapper[4824]: I1124 14:14:36.573483 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:14:40 crc kubenswrapper[4824]: I1124 14:14:40.788444 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:14:40 crc kubenswrapper[4824]: I1124 14:14:40.789039 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:14:42 crc kubenswrapper[4824]: I1124 14:14:42.635838 4824 generic.go:334] "Generic (PLEG): container finished" podID="fe27b019-f6cf-4eec-83df-ef23b6483cf8" containerID="2cec6d9fb80127c3695381979395360b9fb7d24eb420c176be55777dde7f480c" exitCode=0 Nov 24 14:14:42 crc kubenswrapper[4824]: I1124 14:14:42.636239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nh4bn" event={"ID":"fe27b019-f6cf-4eec-83df-ef23b6483cf8","Type":"ContainerDied","Data":"2cec6d9fb80127c3695381979395360b9fb7d24eb420c176be55777dde7f480c"} Nov 24 14:14:43 crc kubenswrapper[4824]: I1124 14:14:43.647535 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nh4bn" event={"ID":"fe27b019-f6cf-4eec-83df-ef23b6483cf8","Type":"ContainerStarted","Data":"bbce604001114340ab576806c441306d36177a7a31b2287a29c7b2765717b267"} Nov 24 14:14:43 crc kubenswrapper[4824]: I1124 14:14:43.672401 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nh4bn" podStartSLOduration=2.103526324 podStartE2EDuration="8.672383155s" podCreationTimestamp="2025-11-24 14:14:35 +0000 UTC" firstStartedPulling="2025-11-24 14:14:36.573267605 +0000 UTC m=+3618.212806905" lastFinishedPulling="2025-11-24 14:14:43.142124426 +0000 UTC m=+3624.781663736" observedRunningTime="2025-11-24 14:14:43.666794888 +0000 UTC m=+3625.306334198" watchObservedRunningTime="2025-11-24 14:14:43.672383155 +0000 UTC m=+3625.311922465" Nov 24 14:14:45 crc kubenswrapper[4824]: I1124 14:14:45.646693 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:45 crc kubenswrapper[4824]: I1124 14:14:45.647204 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:45 crc kubenswrapper[4824]: I1124 14:14:45.714392 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:55 crc kubenswrapper[4824]: I1124 14:14:55.694912 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nh4bn" Nov 24 14:14:55 crc kubenswrapper[4824]: I1124 14:14:55.802945 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nh4bn"] Nov 24 14:14:55 crc kubenswrapper[4824]: I1124 14:14:55.839378 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7xskl"] Nov 24 14:14:55 crc kubenswrapper[4824]: I1124 14:14:55.839868 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7xskl" podUID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerName="registry-server" containerID="cri-o://5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10" gracePeriod=2 Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.470866 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xskl" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.619325 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbjvv\" (UniqueName: \"kubernetes.io/projected/5488c504-7399-4f08-bcb7-d01a64de5da5-kube-api-access-vbjvv\") pod \"5488c504-7399-4f08-bcb7-d01a64de5da5\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.619436 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-catalog-content\") pod \"5488c504-7399-4f08-bcb7-d01a64de5da5\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.619526 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-utilities\") pod \"5488c504-7399-4f08-bcb7-d01a64de5da5\" (UID: \"5488c504-7399-4f08-bcb7-d01a64de5da5\") " Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.623769 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-utilities" (OuterVolumeSpecName: "utilities") pod "5488c504-7399-4f08-bcb7-d01a64de5da5" (UID: "5488c504-7399-4f08-bcb7-d01a64de5da5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.645866 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5488c504-7399-4f08-bcb7-d01a64de5da5-kube-api-access-vbjvv" (OuterVolumeSpecName: "kube-api-access-vbjvv") pod "5488c504-7399-4f08-bcb7-d01a64de5da5" (UID: "5488c504-7399-4f08-bcb7-d01a64de5da5"). InnerVolumeSpecName "kube-api-access-vbjvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.721675 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.721724 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbjvv\" (UniqueName: \"kubernetes.io/projected/5488c504-7399-4f08-bcb7-d01a64de5da5-kube-api-access-vbjvv\") on node \"crc\" DevicePath \"\"" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.731001 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5488c504-7399-4f08-bcb7-d01a64de5da5" (UID: "5488c504-7399-4f08-bcb7-d01a64de5da5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.820423 4824 generic.go:334] "Generic (PLEG): container finished" podID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerID="5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10" exitCode=0 Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.820462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xskl" event={"ID":"5488c504-7399-4f08-bcb7-d01a64de5da5","Type":"ContainerDied","Data":"5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10"} Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.820494 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xskl" event={"ID":"5488c504-7399-4f08-bcb7-d01a64de5da5","Type":"ContainerDied","Data":"d9dffdccac55798d874da23933801070aac5ef8918c6c668545170345cba171e"} Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.820487 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xskl" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.820513 4824 scope.go:117] "RemoveContainer" containerID="5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.823145 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5488c504-7399-4f08-bcb7-d01a64de5da5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.856656 4824 scope.go:117] "RemoveContainer" containerID="f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.861129 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7xskl"] Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.869127 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7xskl"] Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.896644 4824 scope.go:117] "RemoveContainer" containerID="a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.956139 4824 scope.go:117] "RemoveContainer" containerID="5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10" Nov 24 14:14:56 crc kubenswrapper[4824]: E1124 14:14:56.956629 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10\": container with ID starting with 5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10 not found: ID does not exist" containerID="5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.956672 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10"} err="failed to get container status \"5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10\": rpc error: code = NotFound desc = could not find container \"5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10\": container with ID starting with 5f5bda2d24d0837967e937231ceca5c29361f97b747df0420020b61c1c146b10 not found: ID does not exist" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.956700 4824 scope.go:117] "RemoveContainer" containerID="f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd" Nov 24 14:14:56 crc kubenswrapper[4824]: E1124 14:14:56.957072 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd\": container with ID starting with f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd not found: ID does not exist" containerID="f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.957123 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd"} err="failed to get container status \"f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd\": rpc error: code = NotFound desc = could not find container \"f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd\": container with ID starting with f5d511ede759da5c48201885a5f8e91fe0aca00ffe60bd13d43d6d700668d1cd not found: ID does not exist" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.957154 4824 scope.go:117] "RemoveContainer" containerID="a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283" Nov 24 14:14:56 crc kubenswrapper[4824]: E1124 14:14:56.957402 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283\": container with ID starting with a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283 not found: ID does not exist" containerID="a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283" Nov 24 14:14:56 crc kubenswrapper[4824]: I1124 14:14:56.957434 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283"} err="failed to get container status \"a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283\": rpc error: code = NotFound desc = could not find container \"a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283\": container with ID starting with a8ad671e68b1a432f9ad3447aad1bdde5b69b843c58f2acd4f82724c76c77283 not found: ID does not exist" Nov 24 14:14:57 crc kubenswrapper[4824]: I1124 14:14:57.020984 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5488c504-7399-4f08-bcb7-d01a64de5da5" path="/var/lib/kubelet/pods/5488c504-7399-4f08-bcb7-d01a64de5da5/volumes" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.540468 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jv72r"] Nov 24 14:14:58 crc kubenswrapper[4824]: E1124 14:14:58.541229 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerName="extract-utilities" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.541246 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerName="extract-utilities" Nov 24 14:14:58 crc kubenswrapper[4824]: E1124 14:14:58.541265 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerName="extract-content" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.541273 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerName="extract-content" Nov 24 14:14:58 crc kubenswrapper[4824]: E1124 14:14:58.541293 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerName="registry-server" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.541300 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerName="registry-server" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.541537 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5488c504-7399-4f08-bcb7-d01a64de5da5" containerName="registry-server" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.543150 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.551213 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-utilities\") pod \"certified-operators-jv72r\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.551290 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfn6r\" (UniqueName: \"kubernetes.io/projected/37d87012-8ad4-4878-abbf-f8817d9a74f7-kube-api-access-mfn6r\") pod \"certified-operators-jv72r\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.551401 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-catalog-content\") pod \"certified-operators-jv72r\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.568468 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jv72r"] Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.653692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-catalog-content\") pod \"certified-operators-jv72r\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.653784 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-utilities\") pod \"certified-operators-jv72r\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.653846 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfn6r\" (UniqueName: \"kubernetes.io/projected/37d87012-8ad4-4878-abbf-f8817d9a74f7-kube-api-access-mfn6r\") pod \"certified-operators-jv72r\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.654221 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-catalog-content\") pod \"certified-operators-jv72r\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.654372 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-utilities\") pod \"certified-operators-jv72r\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.672621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfn6r\" (UniqueName: \"kubernetes.io/projected/37d87012-8ad4-4878-abbf-f8817d9a74f7-kube-api-access-mfn6r\") pod \"certified-operators-jv72r\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:58 crc kubenswrapper[4824]: I1124 14:14:58.859328 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:14:59 crc kubenswrapper[4824]: I1124 14:14:59.392559 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jv72r"] Nov 24 14:14:59 crc kubenswrapper[4824]: W1124 14:14:59.393929 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37d87012_8ad4_4878_abbf_f8817d9a74f7.slice/crio-d208756748c0f511a29e2bfcfe9813da6fbace3ddd4da17da19fa9197c9f4477 WatchSource:0}: Error finding container d208756748c0f511a29e2bfcfe9813da6fbace3ddd4da17da19fa9197c9f4477: Status 404 returned error can't find the container with id d208756748c0f511a29e2bfcfe9813da6fbace3ddd4da17da19fa9197c9f4477 Nov 24 14:14:59 crc kubenswrapper[4824]: I1124 14:14:59.846083 4824 generic.go:334] "Generic (PLEG): container finished" podID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerID="0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a" exitCode=0 Nov 24 14:14:59 crc kubenswrapper[4824]: I1124 14:14:59.846163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jv72r" event={"ID":"37d87012-8ad4-4878-abbf-f8817d9a74f7","Type":"ContainerDied","Data":"0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a"} Nov 24 14:14:59 crc kubenswrapper[4824]: I1124 14:14:59.846358 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jv72r" event={"ID":"37d87012-8ad4-4878-abbf-f8817d9a74f7","Type":"ContainerStarted","Data":"d208756748c0f511a29e2bfcfe9813da6fbace3ddd4da17da19fa9197c9f4477"} Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.208537 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk"] Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.209776 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.216870 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.217292 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.221583 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk"] Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.282336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ac7bb49-f53a-4d38-a099-412558b4a4d0-config-volume\") pod \"collect-profiles-29399895-httwk\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.282416 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ac7bb49-f53a-4d38-a099-412558b4a4d0-secret-volume\") pod \"collect-profiles-29399895-httwk\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.282580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd4vg\" (UniqueName: \"kubernetes.io/projected/0ac7bb49-f53a-4d38-a099-412558b4a4d0-kube-api-access-rd4vg\") pod \"collect-profiles-29399895-httwk\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.384474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ac7bb49-f53a-4d38-a099-412558b4a4d0-secret-volume\") pod \"collect-profiles-29399895-httwk\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.384880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd4vg\" (UniqueName: \"kubernetes.io/projected/0ac7bb49-f53a-4d38-a099-412558b4a4d0-kube-api-access-rd4vg\") pod \"collect-profiles-29399895-httwk\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.384967 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ac7bb49-f53a-4d38-a099-412558b4a4d0-config-volume\") pod \"collect-profiles-29399895-httwk\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.385984 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ac7bb49-f53a-4d38-a099-412558b4a4d0-config-volume\") pod \"collect-profiles-29399895-httwk\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.393654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ac7bb49-f53a-4d38-a099-412558b4a4d0-secret-volume\") pod \"collect-profiles-29399895-httwk\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.405016 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd4vg\" (UniqueName: \"kubernetes.io/projected/0ac7bb49-f53a-4d38-a099-412558b4a4d0-kube-api-access-rd4vg\") pod \"collect-profiles-29399895-httwk\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:00 crc kubenswrapper[4824]: I1124 14:15:00.540565 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:01 crc kubenswrapper[4824]: I1124 14:15:01.035249 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk"] Nov 24 14:15:01 crc kubenswrapper[4824]: I1124 14:15:01.863405 4824 generic.go:334] "Generic (PLEG): container finished" podID="0ac7bb49-f53a-4d38-a099-412558b4a4d0" containerID="3433425804ec56eaf5207b5c5e29f6800dd25189c9e70b39e00c48652b30d603" exitCode=0 Nov 24 14:15:01 crc kubenswrapper[4824]: I1124 14:15:01.863564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" event={"ID":"0ac7bb49-f53a-4d38-a099-412558b4a4d0","Type":"ContainerDied","Data":"3433425804ec56eaf5207b5c5e29f6800dd25189c9e70b39e00c48652b30d603"} Nov 24 14:15:01 crc kubenswrapper[4824]: I1124 14:15:01.863712 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" event={"ID":"0ac7bb49-f53a-4d38-a099-412558b4a4d0","Type":"ContainerStarted","Data":"93986e88faecd04ed780fe40a7959cd381b947baa996e425566c2566a925097e"} Nov 24 14:15:01 crc kubenswrapper[4824]: I1124 14:15:01.865516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jv72r" event={"ID":"37d87012-8ad4-4878-abbf-f8817d9a74f7","Type":"ContainerStarted","Data":"c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec"} Nov 24 14:15:02 crc kubenswrapper[4824]: I1124 14:15:02.875162 4824 generic.go:334] "Generic (PLEG): container finished" podID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerID="c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec" exitCode=0 Nov 24 14:15:02 crc kubenswrapper[4824]: I1124 14:15:02.876597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jv72r" event={"ID":"37d87012-8ad4-4878-abbf-f8817d9a74f7","Type":"ContainerDied","Data":"c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec"} Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.307475 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.445646 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd4vg\" (UniqueName: \"kubernetes.io/projected/0ac7bb49-f53a-4d38-a099-412558b4a4d0-kube-api-access-rd4vg\") pod \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.446004 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ac7bb49-f53a-4d38-a099-412558b4a4d0-config-volume\") pod \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.446065 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ac7bb49-f53a-4d38-a099-412558b4a4d0-secret-volume\") pod \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\" (UID: \"0ac7bb49-f53a-4d38-a099-412558b4a4d0\") " Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.448117 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ac7bb49-f53a-4d38-a099-412558b4a4d0-config-volume" (OuterVolumeSpecName: "config-volume") pod "0ac7bb49-f53a-4d38-a099-412558b4a4d0" (UID: "0ac7bb49-f53a-4d38-a099-412558b4a4d0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.456085 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac7bb49-f53a-4d38-a099-412558b4a4d0-kube-api-access-rd4vg" (OuterVolumeSpecName: "kube-api-access-rd4vg") pod "0ac7bb49-f53a-4d38-a099-412558b4a4d0" (UID: "0ac7bb49-f53a-4d38-a099-412558b4a4d0"). InnerVolumeSpecName "kube-api-access-rd4vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.457214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac7bb49-f53a-4d38-a099-412558b4a4d0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0ac7bb49-f53a-4d38-a099-412558b4a4d0" (UID: "0ac7bb49-f53a-4d38-a099-412558b4a4d0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.548457 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd4vg\" (UniqueName: \"kubernetes.io/projected/0ac7bb49-f53a-4d38-a099-412558b4a4d0-kube-api-access-rd4vg\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.548498 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ac7bb49-f53a-4d38-a099-412558b4a4d0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.548509 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ac7bb49-f53a-4d38-a099-412558b4a4d0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.888370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jv72r" event={"ID":"37d87012-8ad4-4878-abbf-f8817d9a74f7","Type":"ContainerStarted","Data":"38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c"} Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.891714 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" event={"ID":"0ac7bb49-f53a-4d38-a099-412558b4a4d0","Type":"ContainerDied","Data":"93986e88faecd04ed780fe40a7959cd381b947baa996e425566c2566a925097e"} Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.891748 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93986e88faecd04ed780fe40a7959cd381b947baa996e425566c2566a925097e" Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.891767 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-httwk" Nov 24 14:15:03 crc kubenswrapper[4824]: I1124 14:15:03.915577 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jv72r" podStartSLOduration=2.447193512 podStartE2EDuration="5.915556881s" podCreationTimestamp="2025-11-24 14:14:58 +0000 UTC" firstStartedPulling="2025-11-24 14:14:59.84865038 +0000 UTC m=+3641.488189690" lastFinishedPulling="2025-11-24 14:15:03.317013729 +0000 UTC m=+3644.956553059" observedRunningTime="2025-11-24 14:15:03.904331776 +0000 UTC m=+3645.543871106" watchObservedRunningTime="2025-11-24 14:15:03.915556881 +0000 UTC m=+3645.555096191" Nov 24 14:15:04 crc kubenswrapper[4824]: I1124 14:15:04.390116 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c"] Nov 24 14:15:04 crc kubenswrapper[4824]: I1124 14:15:04.396913 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399850-tmf7c"] Nov 24 14:15:05 crc kubenswrapper[4824]: I1124 14:15:05.032564 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="007fcfe8-9560-471e-9cd0-c4821bffa2d4" path="/var/lib/kubelet/pods/007fcfe8-9560-471e-9cd0-c4821bffa2d4/volumes" Nov 24 14:15:08 crc kubenswrapper[4824]: I1124 14:15:08.860012 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:15:08 crc kubenswrapper[4824]: I1124 14:15:08.860465 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:15:08 crc kubenswrapper[4824]: I1124 14:15:08.913018 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:15:08 crc kubenswrapper[4824]: I1124 14:15:08.999906 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:15:09 crc kubenswrapper[4824]: I1124 14:15:09.153479 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jv72r"] Nov 24 14:15:10 crc kubenswrapper[4824]: I1124 14:15:10.787609 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:15:10 crc kubenswrapper[4824]: I1124 14:15:10.788959 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:15:10 crc kubenswrapper[4824]: I1124 14:15:10.789125 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 14:15:10 crc kubenswrapper[4824]: I1124 14:15:10.790005 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e0b99c526a62db8017cf1ab4e1d16ce7faea62b4707845d45ae09b4bd27370f"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:15:10 crc kubenswrapper[4824]: I1124 14:15:10.790177 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://4e0b99c526a62db8017cf1ab4e1d16ce7faea62b4707845d45ae09b4bd27370f" gracePeriod=600 Nov 24 14:15:10 crc kubenswrapper[4824]: I1124 14:15:10.985357 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="4e0b99c526a62db8017cf1ab4e1d16ce7faea62b4707845d45ae09b4bd27370f" exitCode=0 Nov 24 14:15:10 crc kubenswrapper[4824]: I1124 14:15:10.985576 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jv72r" podUID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerName="registry-server" containerID="cri-o://38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c" gracePeriod=2 Nov 24 14:15:10 crc kubenswrapper[4824]: I1124 14:15:10.985668 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"4e0b99c526a62db8017cf1ab4e1d16ce7faea62b4707845d45ae09b4bd27370f"} Nov 24 14:15:10 crc kubenswrapper[4824]: I1124 14:15:10.985698 4824 scope.go:117] "RemoveContainer" containerID="f05344fdedc177f6c3153ed35c906dff646e7a4f774c2efd62e188905879483c" Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.602471 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.710677 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfn6r\" (UniqueName: \"kubernetes.io/projected/37d87012-8ad4-4878-abbf-f8817d9a74f7-kube-api-access-mfn6r\") pod \"37d87012-8ad4-4878-abbf-f8817d9a74f7\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.710846 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-utilities\") pod \"37d87012-8ad4-4878-abbf-f8817d9a74f7\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.710895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-catalog-content\") pod \"37d87012-8ad4-4878-abbf-f8817d9a74f7\" (UID: \"37d87012-8ad4-4878-abbf-f8817d9a74f7\") " Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.711670 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-utilities" (OuterVolumeSpecName: "utilities") pod "37d87012-8ad4-4878-abbf-f8817d9a74f7" (UID: "37d87012-8ad4-4878-abbf-f8817d9a74f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.717029 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37d87012-8ad4-4878-abbf-f8817d9a74f7-kube-api-access-mfn6r" (OuterVolumeSpecName: "kube-api-access-mfn6r") pod "37d87012-8ad4-4878-abbf-f8817d9a74f7" (UID: "37d87012-8ad4-4878-abbf-f8817d9a74f7"). InnerVolumeSpecName "kube-api-access-mfn6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.761835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37d87012-8ad4-4878-abbf-f8817d9a74f7" (UID: "37d87012-8ad4-4878-abbf-f8817d9a74f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.812970 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfn6r\" (UniqueName: \"kubernetes.io/projected/37d87012-8ad4-4878-abbf-f8817d9a74f7-kube-api-access-mfn6r\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.813004 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.813015 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d87012-8ad4-4878-abbf-f8817d9a74f7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.995835 4824 generic.go:334] "Generic (PLEG): container finished" podID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerID="38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c" exitCode=0 Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.995914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jv72r" event={"ID":"37d87012-8ad4-4878-abbf-f8817d9a74f7","Type":"ContainerDied","Data":"38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c"} Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.995946 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jv72r" event={"ID":"37d87012-8ad4-4878-abbf-f8817d9a74f7","Type":"ContainerDied","Data":"d208756748c0f511a29e2bfcfe9813da6fbace3ddd4da17da19fa9197c9f4477"} Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.995991 4824 scope.go:117] "RemoveContainer" containerID="38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c" Nov 24 14:15:11 crc kubenswrapper[4824]: I1124 14:15:11.996083 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jv72r" Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.002733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3"} Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.037040 4824 scope.go:117] "RemoveContainer" containerID="c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec" Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.051924 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jv72r"] Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.059458 4824 scope.go:117] "RemoveContainer" containerID="0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a" Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.062543 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jv72r"] Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.115994 4824 scope.go:117] "RemoveContainer" containerID="38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c" Nov 24 14:15:12 crc kubenswrapper[4824]: E1124 14:15:12.116510 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c\": container with ID starting with 38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c not found: ID does not exist" containerID="38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c" Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.116544 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c"} err="failed to get container status \"38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c\": rpc error: code = NotFound desc = could not find container \"38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c\": container with ID starting with 38e0aa6e49cac6e964f40ecd11318b449f3f32250e24798534176e68cb6b9c7c not found: ID does not exist" Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.116565 4824 scope.go:117] "RemoveContainer" containerID="c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec" Nov 24 14:15:12 crc kubenswrapper[4824]: E1124 14:15:12.116920 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec\": container with ID starting with c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec not found: ID does not exist" containerID="c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec" Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.116947 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec"} err="failed to get container status \"c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec\": rpc error: code = NotFound desc = could not find container \"c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec\": container with ID starting with c7be88052b7046fdf4dc180ae78066ba4b5b9cabaadb559358ce1ff1c9d74bec not found: ID does not exist" Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.116962 4824 scope.go:117] "RemoveContainer" containerID="0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a" Nov 24 14:15:12 crc kubenswrapper[4824]: E1124 14:15:12.117248 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a\": container with ID starting with 0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a not found: ID does not exist" containerID="0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a" Nov 24 14:15:12 crc kubenswrapper[4824]: I1124 14:15:12.117268 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a"} err="failed to get container status \"0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a\": rpc error: code = NotFound desc = could not find container \"0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a\": container with ID starting with 0e219d22af300b3b37282a3eafbe20b38fc78b5ba89e077ed2d88b8c52840b0a not found: ID does not exist" Nov 24 14:15:13 crc kubenswrapper[4824]: I1124 14:15:13.024364 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37d87012-8ad4-4878-abbf-f8817d9a74f7" path="/var/lib/kubelet/pods/37d87012-8ad4-4878-abbf-f8817d9a74f7/volumes" Nov 24 14:15:25 crc kubenswrapper[4824]: I1124 14:15:25.075472 4824 scope.go:117] "RemoveContainer" containerID="4a7d9bfbb6bc4d3d2ed940525a1df1d99010a102ca5c61ca90d0d5c7d11c683e" Nov 24 14:17:40 crc kubenswrapper[4824]: I1124 14:17:40.787765 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:17:40 crc kubenswrapper[4824]: I1124 14:17:40.788311 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:18:10 crc kubenswrapper[4824]: I1124 14:18:10.787547 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:18:10 crc kubenswrapper[4824]: I1124 14:18:10.788567 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:18:40 crc kubenswrapper[4824]: I1124 14:18:40.787755 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:18:40 crc kubenswrapper[4824]: I1124 14:18:40.788339 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:18:40 crc kubenswrapper[4824]: I1124 14:18:40.788381 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 14:18:40 crc kubenswrapper[4824]: I1124 14:18:40.789416 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:18:40 crc kubenswrapper[4824]: I1124 14:18:40.789555 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" gracePeriod=600 Nov 24 14:18:40 crc kubenswrapper[4824]: E1124 14:18:40.915454 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:18:41 crc kubenswrapper[4824]: I1124 14:18:41.018742 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" exitCode=0 Nov 24 14:18:41 crc kubenswrapper[4824]: I1124 14:18:41.023348 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3"} Nov 24 14:18:41 crc kubenswrapper[4824]: I1124 14:18:41.023416 4824 scope.go:117] "RemoveContainer" containerID="4e0b99c526a62db8017cf1ab4e1d16ce7faea62b4707845d45ae09b4bd27370f" Nov 24 14:18:41 crc kubenswrapper[4824]: I1124 14:18:41.024157 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:18:41 crc kubenswrapper[4824]: E1124 14:18:41.024454 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.772258 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sx72p"] Nov 24 14:18:51 crc kubenswrapper[4824]: E1124 14:18:51.773223 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7bb49-f53a-4d38-a099-412558b4a4d0" containerName="collect-profiles" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.773239 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7bb49-f53a-4d38-a099-412558b4a4d0" containerName="collect-profiles" Nov 24 14:18:51 crc kubenswrapper[4824]: E1124 14:18:51.773250 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerName="extract-utilities" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.773260 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerName="extract-utilities" Nov 24 14:18:51 crc kubenswrapper[4824]: E1124 14:18:51.773298 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerName="extract-content" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.773306 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerName="extract-content" Nov 24 14:18:51 crc kubenswrapper[4824]: E1124 14:18:51.773325 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerName="registry-server" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.773332 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerName="registry-server" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.773596 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="37d87012-8ad4-4878-abbf-f8817d9a74f7" containerName="registry-server" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.773609 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac7bb49-f53a-4d38-a099-412558b4a4d0" containerName="collect-profiles" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.775291 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.794316 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sx72p"] Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.951259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4ppg\" (UniqueName: \"kubernetes.io/projected/a70f90ce-229f-4334-8458-8eb12bf7d4cc-kube-api-access-k4ppg\") pod \"redhat-operators-sx72p\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.951344 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-catalog-content\") pod \"redhat-operators-sx72p\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:51 crc kubenswrapper[4824]: I1124 14:18:51.951452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-utilities\") pod \"redhat-operators-sx72p\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:52 crc kubenswrapper[4824]: I1124 14:18:52.010227 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:18:52 crc kubenswrapper[4824]: E1124 14:18:52.010542 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:18:52 crc kubenswrapper[4824]: I1124 14:18:52.052748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4ppg\" (UniqueName: \"kubernetes.io/projected/a70f90ce-229f-4334-8458-8eb12bf7d4cc-kube-api-access-k4ppg\") pod \"redhat-operators-sx72p\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:52 crc kubenswrapper[4824]: I1124 14:18:52.052852 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-catalog-content\") pod \"redhat-operators-sx72p\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:52 crc kubenswrapper[4824]: I1124 14:18:52.052926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-utilities\") pod \"redhat-operators-sx72p\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:52 crc kubenswrapper[4824]: I1124 14:18:52.053435 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-utilities\") pod \"redhat-operators-sx72p\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:52 crc kubenswrapper[4824]: I1124 14:18:52.053507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-catalog-content\") pod \"redhat-operators-sx72p\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:52 crc kubenswrapper[4824]: I1124 14:18:52.089946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4ppg\" (UniqueName: \"kubernetes.io/projected/a70f90ce-229f-4334-8458-8eb12bf7d4cc-kube-api-access-k4ppg\") pod \"redhat-operators-sx72p\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:52 crc kubenswrapper[4824]: I1124 14:18:52.115637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:18:52 crc kubenswrapper[4824]: I1124 14:18:52.662302 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sx72p"] Nov 24 14:18:53 crc kubenswrapper[4824]: I1124 14:18:53.143072 4824 generic.go:334] "Generic (PLEG): container finished" podID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerID="78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62" exitCode=0 Nov 24 14:18:53 crc kubenswrapper[4824]: I1124 14:18:53.143223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx72p" event={"ID":"a70f90ce-229f-4334-8458-8eb12bf7d4cc","Type":"ContainerDied","Data":"78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62"} Nov 24 14:18:53 crc kubenswrapper[4824]: I1124 14:18:53.143473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx72p" event={"ID":"a70f90ce-229f-4334-8458-8eb12bf7d4cc","Type":"ContainerStarted","Data":"19b063f8950e16aad7b56ba7b2c7d14a35172d44cb3fe4f38e056d0d19b3df0f"} Nov 24 14:18:55 crc kubenswrapper[4824]: I1124 14:18:55.171360 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx72p" event={"ID":"a70f90ce-229f-4334-8458-8eb12bf7d4cc","Type":"ContainerStarted","Data":"5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e"} Nov 24 14:19:03 crc kubenswrapper[4824]: I1124 14:19:03.251100 4824 generic.go:334] "Generic (PLEG): container finished" podID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerID="5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e" exitCode=0 Nov 24 14:19:03 crc kubenswrapper[4824]: I1124 14:19:03.251697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx72p" event={"ID":"a70f90ce-229f-4334-8458-8eb12bf7d4cc","Type":"ContainerDied","Data":"5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e"} Nov 24 14:19:04 crc kubenswrapper[4824]: I1124 14:19:04.264973 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx72p" event={"ID":"a70f90ce-229f-4334-8458-8eb12bf7d4cc","Type":"ContainerStarted","Data":"a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e"} Nov 24 14:19:07 crc kubenswrapper[4824]: I1124 14:19:07.010899 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:19:07 crc kubenswrapper[4824]: E1124 14:19:07.011478 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:19:12 crc kubenswrapper[4824]: I1124 14:19:12.116116 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:19:12 crc kubenswrapper[4824]: I1124 14:19:12.116694 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:19:13 crc kubenswrapper[4824]: I1124 14:19:13.168090 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sx72p" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="registry-server" probeResult="failure" output=< Nov 24 14:19:13 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:19:13 crc kubenswrapper[4824]: > Nov 24 14:19:21 crc kubenswrapper[4824]: I1124 14:19:21.009883 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:19:21 crc kubenswrapper[4824]: E1124 14:19:21.010525 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:19:23 crc kubenswrapper[4824]: I1124 14:19:23.179654 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sx72p" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="registry-server" probeResult="failure" output=< Nov 24 14:19:23 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:19:23 crc kubenswrapper[4824]: > Nov 24 14:19:32 crc kubenswrapper[4824]: I1124 14:19:32.010408 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:19:32 crc kubenswrapper[4824]: E1124 14:19:32.011182 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:19:33 crc kubenswrapper[4824]: I1124 14:19:33.164877 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sx72p" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="registry-server" probeResult="failure" output=< Nov 24 14:19:33 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:19:33 crc kubenswrapper[4824]: > Nov 24 14:19:43 crc kubenswrapper[4824]: I1124 14:19:43.182736 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sx72p" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="registry-server" probeResult="failure" output=< Nov 24 14:19:43 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:19:43 crc kubenswrapper[4824]: > Nov 24 14:19:45 crc kubenswrapper[4824]: I1124 14:19:45.010321 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:19:45 crc kubenswrapper[4824]: E1124 14:19:45.010909 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:19:52 crc kubenswrapper[4824]: I1124 14:19:52.190594 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:19:52 crc kubenswrapper[4824]: I1124 14:19:52.215274 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sx72p" podStartSLOduration=50.540609162 podStartE2EDuration="1m1.215251475s" podCreationTimestamp="2025-11-24 14:18:51 +0000 UTC" firstStartedPulling="2025-11-24 14:18:53.147151722 +0000 UTC m=+3874.786691032" lastFinishedPulling="2025-11-24 14:19:03.821794035 +0000 UTC m=+3885.461333345" observedRunningTime="2025-11-24 14:19:04.294669368 +0000 UTC m=+3885.934208678" watchObservedRunningTime="2025-11-24 14:19:52.215251475 +0000 UTC m=+3933.854790785" Nov 24 14:19:52 crc kubenswrapper[4824]: I1124 14:19:52.256796 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:19:53 crc kubenswrapper[4824]: I1124 14:19:53.002489 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sx72p"] Nov 24 14:19:53 crc kubenswrapper[4824]: I1124 14:19:53.704131 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sx72p" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="registry-server" containerID="cri-o://a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e" gracePeriod=2 Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.303496 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.431357 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-utilities\") pod \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.431447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4ppg\" (UniqueName: \"kubernetes.io/projected/a70f90ce-229f-4334-8458-8eb12bf7d4cc-kube-api-access-k4ppg\") pod \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.431552 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-catalog-content\") pod \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\" (UID: \"a70f90ce-229f-4334-8458-8eb12bf7d4cc\") " Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.432746 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-utilities" (OuterVolumeSpecName: "utilities") pod "a70f90ce-229f-4334-8458-8eb12bf7d4cc" (UID: "a70f90ce-229f-4334-8458-8eb12bf7d4cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.438021 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70f90ce-229f-4334-8458-8eb12bf7d4cc-kube-api-access-k4ppg" (OuterVolumeSpecName: "kube-api-access-k4ppg") pod "a70f90ce-229f-4334-8458-8eb12bf7d4cc" (UID: "a70f90ce-229f-4334-8458-8eb12bf7d4cc"). InnerVolumeSpecName "kube-api-access-k4ppg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.533949 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.533988 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4ppg\" (UniqueName: \"kubernetes.io/projected/a70f90ce-229f-4334-8458-8eb12bf7d4cc-kube-api-access-k4ppg\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.547039 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a70f90ce-229f-4334-8458-8eb12bf7d4cc" (UID: "a70f90ce-229f-4334-8458-8eb12bf7d4cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.637328 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70f90ce-229f-4334-8458-8eb12bf7d4cc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.716864 4824 generic.go:334] "Generic (PLEG): container finished" podID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerID="a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e" exitCode=0 Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.716929 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx72p" event={"ID":"a70f90ce-229f-4334-8458-8eb12bf7d4cc","Type":"ContainerDied","Data":"a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e"} Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.716966 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx72p" event={"ID":"a70f90ce-229f-4334-8458-8eb12bf7d4cc","Type":"ContainerDied","Data":"19b063f8950e16aad7b56ba7b2c7d14a35172d44cb3fe4f38e056d0d19b3df0f"} Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.716993 4824 scope.go:117] "RemoveContainer" containerID="a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.717185 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx72p" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.760662 4824 scope.go:117] "RemoveContainer" containerID="5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.777435 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sx72p"] Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.808102 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sx72p"] Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.811607 4824 scope.go:117] "RemoveContainer" containerID="78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.866337 4824 scope.go:117] "RemoveContainer" containerID="a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e" Nov 24 14:19:54 crc kubenswrapper[4824]: E1124 14:19:54.866862 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e\": container with ID starting with a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e not found: ID does not exist" containerID="a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.866912 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e"} err="failed to get container status \"a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e\": rpc error: code = NotFound desc = could not find container \"a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e\": container with ID starting with a467487e8fcdede98163eb35706eab7cdcc7600c94aecc2d2680982d57daa83e not found: ID does not exist" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.866941 4824 scope.go:117] "RemoveContainer" containerID="5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e" Nov 24 14:19:54 crc kubenswrapper[4824]: E1124 14:19:54.870773 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e\": container with ID starting with 5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e not found: ID does not exist" containerID="5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.870851 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e"} err="failed to get container status \"5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e\": rpc error: code = NotFound desc = could not find container \"5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e\": container with ID starting with 5bfe4747fe8d360e958eea9f6ac441239362e2516734ba3ad1bce0958dd8259e not found: ID does not exist" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.870885 4824 scope.go:117] "RemoveContainer" containerID="78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62" Nov 24 14:19:54 crc kubenswrapper[4824]: E1124 14:19:54.871543 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62\": container with ID starting with 78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62 not found: ID does not exist" containerID="78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62" Nov 24 14:19:54 crc kubenswrapper[4824]: I1124 14:19:54.871566 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62"} err="failed to get container status \"78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62\": rpc error: code = NotFound desc = could not find container \"78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62\": container with ID starting with 78c73803988e7a6bedcd9944bf9f6bba98dae3aba90263613320238cda5b5b62 not found: ID does not exist" Nov 24 14:19:55 crc kubenswrapper[4824]: I1124 14:19:55.029006 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" path="/var/lib/kubelet/pods/a70f90ce-229f-4334-8458-8eb12bf7d4cc/volumes" Nov 24 14:19:59 crc kubenswrapper[4824]: I1124 14:19:59.018130 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:19:59 crc kubenswrapper[4824]: E1124 14:19:59.019182 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:20:12 crc kubenswrapper[4824]: I1124 14:20:12.010519 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:20:12 crc kubenswrapper[4824]: E1124 14:20:12.011540 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:20:26 crc kubenswrapper[4824]: I1124 14:20:26.010271 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:20:26 crc kubenswrapper[4824]: E1124 14:20:26.010975 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:20:37 crc kubenswrapper[4824]: I1124 14:20:37.013604 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:20:37 crc kubenswrapper[4824]: E1124 14:20:37.015550 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:20:50 crc kubenswrapper[4824]: I1124 14:20:50.010863 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:20:50 crc kubenswrapper[4824]: E1124 14:20:50.011699 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:21:01 crc kubenswrapper[4824]: I1124 14:21:01.010959 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:21:01 crc kubenswrapper[4824]: E1124 14:21:01.011736 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:21:14 crc kubenswrapper[4824]: I1124 14:21:14.010440 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:21:14 crc kubenswrapper[4824]: E1124 14:21:14.011260 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:21:26 crc kubenswrapper[4824]: I1124 14:21:26.010312 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:21:26 crc kubenswrapper[4824]: E1124 14:21:26.011111 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:21:41 crc kubenswrapper[4824]: I1124 14:21:41.009993 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:21:41 crc kubenswrapper[4824]: E1124 14:21:41.010699 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:21:56 crc kubenswrapper[4824]: I1124 14:21:56.010638 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:21:56 crc kubenswrapper[4824]: E1124 14:21:56.011607 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:22:09 crc kubenswrapper[4824]: I1124 14:22:09.013004 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:22:09 crc kubenswrapper[4824]: E1124 14:22:09.014195 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:22:20 crc kubenswrapper[4824]: I1124 14:22:20.011558 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:22:20 crc kubenswrapper[4824]: E1124 14:22:20.012524 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:22:31 crc kubenswrapper[4824]: I1124 14:22:31.238868 4824 generic.go:334] "Generic (PLEG): container finished" podID="a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" containerID="bcd1546a48cc865529bb47c2eaa60d9ed935f25e27bade958e6f2482a2ccf4f0" exitCode=0 Nov 24 14:22:31 crc kubenswrapper[4824]: I1124 14:22:31.238921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0","Type":"ContainerDied","Data":"bcd1546a48cc865529bb47c2eaa60d9ed935f25e27bade958e6f2482a2ccf4f0"} Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.010892 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:22:32 crc kubenswrapper[4824]: E1124 14:22:32.011431 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.714603 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.892387 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ca-certs\") pod \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.892466 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config-secret\") pod \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.892641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-workdir\") pod \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.892712 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx7jt\" (UniqueName: \"kubernetes.io/projected/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-kube-api-access-kx7jt\") pod \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.892745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-temporary\") pod \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.892772 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config\") pod \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.892789 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.892835 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-config-data\") pod \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.892883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ssh-key\") pod \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\" (UID: \"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0\") " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.893590 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" (UID: "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.894790 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-config-data" (OuterVolumeSpecName: "config-data") pod "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" (UID: "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.901737 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" (UID: "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.905509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" (UID: "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.906281 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-kube-api-access-kx7jt" (OuterVolumeSpecName: "kube-api-access-kx7jt") pod "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" (UID: "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0"). InnerVolumeSpecName "kube-api-access-kx7jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.921336 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" (UID: "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.921956 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" (UID: "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.926514 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" (UID: "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.943643 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" (UID: "a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.995161 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.995201 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx7jt\" (UniqueName: \"kubernetes.io/projected/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-kube-api-access-kx7jt\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.995212 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.995223 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.998032 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.998051 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.998060 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.998070 4824 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:32 crc kubenswrapper[4824]: I1124 14:22:32.998079 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:33 crc kubenswrapper[4824]: I1124 14:22:33.018188 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 24 14:22:33 crc kubenswrapper[4824]: I1124 14:22:33.099612 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:22:33 crc kubenswrapper[4824]: I1124 14:22:33.262083 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0","Type":"ContainerDied","Data":"1b89ff540eff077c43467e51a8dfff04ade825414083d67c4890eabb89ce956b"} Nov 24 14:22:33 crc kubenswrapper[4824]: I1124 14:22:33.262123 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b89ff540eff077c43467e51a8dfff04ade825414083d67c4890eabb89ce956b" Nov 24 14:22:33 crc kubenswrapper[4824]: I1124 14:22:33.262562 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.254600 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 14:22:44 crc kubenswrapper[4824]: E1124 14:22:44.255590 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="registry-server" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.255607 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="registry-server" Nov 24 14:22:44 crc kubenswrapper[4824]: E1124 14:22:44.255623 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="extract-utilities" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.255630 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="extract-utilities" Nov 24 14:22:44 crc kubenswrapper[4824]: E1124 14:22:44.255639 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="extract-content" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.255646 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="extract-content" Nov 24 14:22:44 crc kubenswrapper[4824]: E1124 14:22:44.255660 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" containerName="tempest-tests-tempest-tests-runner" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.255667 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" containerName="tempest-tests-tempest-tests-runner" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.255940 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0" containerName="tempest-tests-tempest-tests-runner" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.255957 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a70f90ce-229f-4334-8458-8eb12bf7d4cc" containerName="registry-server" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.256728 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.258833 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qvsng" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.263886 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.408449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkph4\" (UniqueName: \"kubernetes.io/projected/b719b5e7-33d8-43ec-9d61-19bd0676886a-kube-api-access-hkph4\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b719b5e7-33d8-43ec-9d61-19bd0676886a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.408820 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b719b5e7-33d8-43ec-9d61-19bd0676886a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.510226 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkph4\" (UniqueName: \"kubernetes.io/projected/b719b5e7-33d8-43ec-9d61-19bd0676886a-kube-api-access-hkph4\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b719b5e7-33d8-43ec-9d61-19bd0676886a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.510279 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b719b5e7-33d8-43ec-9d61-19bd0676886a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.510763 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b719b5e7-33d8-43ec-9d61-19bd0676886a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.533621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkph4\" (UniqueName: \"kubernetes.io/projected/b719b5e7-33d8-43ec-9d61-19bd0676886a-kube-api-access-hkph4\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b719b5e7-33d8-43ec-9d61-19bd0676886a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.541644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b719b5e7-33d8-43ec-9d61-19bd0676886a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:22:44 crc kubenswrapper[4824]: I1124 14:22:44.580796 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:22:45 crc kubenswrapper[4824]: I1124 14:22:45.027157 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 14:22:45 crc kubenswrapper[4824]: I1124 14:22:45.040282 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:22:45 crc kubenswrapper[4824]: I1124 14:22:45.362479 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b719b5e7-33d8-43ec-9d61-19bd0676886a","Type":"ContainerStarted","Data":"049a11a913f73d1e3339ef15bf3aa6bad84da5c230762176d1fff69fb8a9eba0"} Nov 24 14:22:47 crc kubenswrapper[4824]: I1124 14:22:47.010430 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:22:47 crc kubenswrapper[4824]: E1124 14:22:47.011117 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:22:47 crc kubenswrapper[4824]: I1124 14:22:47.380328 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b719b5e7-33d8-43ec-9d61-19bd0676886a","Type":"ContainerStarted","Data":"1978106d58b02002975af20c5512a6b0eef1e0522383b1cee3eaec43cd41fc54"} Nov 24 14:22:47 crc kubenswrapper[4824]: I1124 14:22:47.402745 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.075211019 podStartE2EDuration="3.402727637s" podCreationTimestamp="2025-11-24 14:22:44 +0000 UTC" firstStartedPulling="2025-11-24 14:22:45.040063246 +0000 UTC m=+4106.679602556" lastFinishedPulling="2025-11-24 14:22:46.367579864 +0000 UTC m=+4108.007119174" observedRunningTime="2025-11-24 14:22:47.394246685 +0000 UTC m=+4109.033785995" watchObservedRunningTime="2025-11-24 14:22:47.402727637 +0000 UTC m=+4109.042266947" Nov 24 14:22:59 crc kubenswrapper[4824]: I1124 14:22:59.019935 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:22:59 crc kubenswrapper[4824]: E1124 14:22:59.022244 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.594525 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hk622/must-gather-dkjtr"] Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.597535 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.601090 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-hk622"/"default-dockercfg-hggb5" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.601260 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hk622"/"openshift-service-ca.crt" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.601482 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hk622"/"kube-root-ca.crt" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.676497 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hk622/must-gather-dkjtr"] Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.723368 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-must-gather-output\") pod \"must-gather-dkjtr\" (UID: \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\") " pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.723640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jjn8\" (UniqueName: \"kubernetes.io/projected/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-kube-api-access-8jjn8\") pod \"must-gather-dkjtr\" (UID: \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\") " pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.826224 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jjn8\" (UniqueName: \"kubernetes.io/projected/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-kube-api-access-8jjn8\") pod \"must-gather-dkjtr\" (UID: \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\") " pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.826707 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-must-gather-output\") pod \"must-gather-dkjtr\" (UID: \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\") " pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.827310 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-must-gather-output\") pod \"must-gather-dkjtr\" (UID: \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\") " pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.850871 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jjn8\" (UniqueName: \"kubernetes.io/projected/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-kube-api-access-8jjn8\") pod \"must-gather-dkjtr\" (UID: \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\") " pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:23:09 crc kubenswrapper[4824]: I1124 14:23:09.913078 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:23:12 crc kubenswrapper[4824]: I1124 14:23:10.529829 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hk622/must-gather-dkjtr"] Nov 24 14:23:12 crc kubenswrapper[4824]: I1124 14:23:10.604157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/must-gather-dkjtr" event={"ID":"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0","Type":"ContainerStarted","Data":"01f79ff05afe90c8466ec4a29a4b2ac3ac8a4105043fb9059c11562db9a96837"} Nov 24 14:23:14 crc kubenswrapper[4824]: I1124 14:23:14.010309 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:23:14 crc kubenswrapper[4824]: E1124 14:23:14.011101 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:23:16 crc kubenswrapper[4824]: I1124 14:23:16.656678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/must-gather-dkjtr" event={"ID":"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0","Type":"ContainerStarted","Data":"036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960"} Nov 24 14:23:16 crc kubenswrapper[4824]: I1124 14:23:16.657121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/must-gather-dkjtr" event={"ID":"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0","Type":"ContainerStarted","Data":"fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1"} Nov 24 14:23:16 crc kubenswrapper[4824]: I1124 14:23:16.679408 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hk622/must-gather-dkjtr" podStartSLOduration=2.547811704 podStartE2EDuration="7.679390321s" podCreationTimestamp="2025-11-24 14:23:09 +0000 UTC" firstStartedPulling="2025-11-24 14:23:10.542943649 +0000 UTC m=+4132.182482959" lastFinishedPulling="2025-11-24 14:23:15.674522246 +0000 UTC m=+4137.314061576" observedRunningTime="2025-11-24 14:23:16.674042027 +0000 UTC m=+4138.313581338" watchObservedRunningTime="2025-11-24 14:23:16.679390321 +0000 UTC m=+4138.318929631" Nov 24 14:23:25 crc kubenswrapper[4824]: I1124 14:23:25.010451 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:23:25 crc kubenswrapper[4824]: E1124 14:23:25.011273 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:23:25 crc kubenswrapper[4824]: E1124 14:23:25.587764 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.188:47588->38.129.56.188:33927: write tcp 38.129.56.188:47588->38.129.56.188:33927: write: broken pipe Nov 24 14:23:26 crc kubenswrapper[4824]: I1124 14:23:26.521710 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hk622/crc-debug-gx82j"] Nov 24 14:23:26 crc kubenswrapper[4824]: I1124 14:23:26.522830 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:23:26 crc kubenswrapper[4824]: I1124 14:23:26.582903 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cd566d4-13a3-423a-894d-d8dcb52325f7-host\") pod \"crc-debug-gx82j\" (UID: \"7cd566d4-13a3-423a-894d-d8dcb52325f7\") " pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:23:26 crc kubenswrapper[4824]: I1124 14:23:26.583306 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzk92\" (UniqueName: \"kubernetes.io/projected/7cd566d4-13a3-423a-894d-d8dcb52325f7-kube-api-access-pzk92\") pod \"crc-debug-gx82j\" (UID: \"7cd566d4-13a3-423a-894d-d8dcb52325f7\") " pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:23:26 crc kubenswrapper[4824]: I1124 14:23:26.685485 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cd566d4-13a3-423a-894d-d8dcb52325f7-host\") pod \"crc-debug-gx82j\" (UID: \"7cd566d4-13a3-423a-894d-d8dcb52325f7\") " pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:23:26 crc kubenswrapper[4824]: I1124 14:23:26.685627 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzk92\" (UniqueName: \"kubernetes.io/projected/7cd566d4-13a3-423a-894d-d8dcb52325f7-kube-api-access-pzk92\") pod \"crc-debug-gx82j\" (UID: \"7cd566d4-13a3-423a-894d-d8dcb52325f7\") " pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:23:26 crc kubenswrapper[4824]: I1124 14:23:26.686216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cd566d4-13a3-423a-894d-d8dcb52325f7-host\") pod \"crc-debug-gx82j\" (UID: \"7cd566d4-13a3-423a-894d-d8dcb52325f7\") " pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:23:26 crc kubenswrapper[4824]: I1124 14:23:26.705322 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzk92\" (UniqueName: \"kubernetes.io/projected/7cd566d4-13a3-423a-894d-d8dcb52325f7-kube-api-access-pzk92\") pod \"crc-debug-gx82j\" (UID: \"7cd566d4-13a3-423a-894d-d8dcb52325f7\") " pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:23:26 crc kubenswrapper[4824]: I1124 14:23:26.841860 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:23:26 crc kubenswrapper[4824]: W1124 14:23:26.873308 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cd566d4_13a3_423a_894d_d8dcb52325f7.slice/crio-2a76de55d06ea5e37bdc25c67122f88d8152e45dd0a5550248f7b97f598993f3 WatchSource:0}: Error finding container 2a76de55d06ea5e37bdc25c67122f88d8152e45dd0a5550248f7b97f598993f3: Status 404 returned error can't find the container with id 2a76de55d06ea5e37bdc25c67122f88d8152e45dd0a5550248f7b97f598993f3 Nov 24 14:23:27 crc kubenswrapper[4824]: I1124 14:23:27.750878 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/crc-debug-gx82j" event={"ID":"7cd566d4-13a3-423a-894d-d8dcb52325f7","Type":"ContainerStarted","Data":"2a76de55d06ea5e37bdc25c67122f88d8152e45dd0a5550248f7b97f598993f3"} Nov 24 14:23:38 crc kubenswrapper[4824]: I1124 14:23:38.923073 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/crc-debug-gx82j" event={"ID":"7cd566d4-13a3-423a-894d-d8dcb52325f7","Type":"ContainerStarted","Data":"278a250b83609700c67e370907958cbc5223716a9a61a77f196c35a24fb14670"} Nov 24 14:23:38 crc kubenswrapper[4824]: I1124 14:23:38.946610 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hk622/crc-debug-gx82j" podStartSLOduration=2.266665227 podStartE2EDuration="12.946591068s" podCreationTimestamp="2025-11-24 14:23:26 +0000 UTC" firstStartedPulling="2025-11-24 14:23:26.876098871 +0000 UTC m=+4148.515638191" lastFinishedPulling="2025-11-24 14:23:37.556024702 +0000 UTC m=+4159.195564032" observedRunningTime="2025-11-24 14:23:38.9462742 +0000 UTC m=+4160.585813530" watchObservedRunningTime="2025-11-24 14:23:38.946591068 +0000 UTC m=+4160.586130378" Nov 24 14:23:40 crc kubenswrapper[4824]: I1124 14:23:40.010042 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:23:40 crc kubenswrapper[4824]: E1124 14:23:40.011190 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:23:53 crc kubenswrapper[4824]: I1124 14:23:53.010093 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:23:53 crc kubenswrapper[4824]: I1124 14:23:53.969609 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-njkn6"] Nov 24 14:23:53 crc kubenswrapper[4824]: I1124 14:23:53.972170 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:53 crc kubenswrapper[4824]: I1124 14:23:53.992215 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-njkn6"] Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.051142 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-catalog-content\") pod \"redhat-marketplace-njkn6\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.051196 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddchj\" (UniqueName: \"kubernetes.io/projected/dface4f6-f5da-409b-879d-a6057f430e79-kube-api-access-ddchj\") pod \"redhat-marketplace-njkn6\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.051259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-utilities\") pod \"redhat-marketplace-njkn6\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.153145 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-utilities\") pod \"redhat-marketplace-njkn6\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.153300 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-catalog-content\") pod \"redhat-marketplace-njkn6\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.153325 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddchj\" (UniqueName: \"kubernetes.io/projected/dface4f6-f5da-409b-879d-a6057f430e79-kube-api-access-ddchj\") pod \"redhat-marketplace-njkn6\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.153703 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-utilities\") pod \"redhat-marketplace-njkn6\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.153702 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-catalog-content\") pod \"redhat-marketplace-njkn6\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.178572 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddchj\" (UniqueName: \"kubernetes.io/projected/dface4f6-f5da-409b-879d-a6057f430e79-kube-api-access-ddchj\") pod \"redhat-marketplace-njkn6\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:54 crc kubenswrapper[4824]: I1124 14:23:54.298123 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:23:57 crc kubenswrapper[4824]: I1124 14:23:57.076582 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"186331f57cfd3cccc17004a62f407141eee208514b6cb45c3ca9fa815ac14b2c"} Nov 24 14:23:57 crc kubenswrapper[4824]: I1124 14:23:57.237514 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-njkn6"] Nov 24 14:23:58 crc kubenswrapper[4824]: I1124 14:23:58.088079 4824 generic.go:334] "Generic (PLEG): container finished" podID="dface4f6-f5da-409b-879d-a6057f430e79" containerID="f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5" exitCode=0 Nov 24 14:23:58 crc kubenswrapper[4824]: I1124 14:23:58.088142 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njkn6" event={"ID":"dface4f6-f5da-409b-879d-a6057f430e79","Type":"ContainerDied","Data":"f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5"} Nov 24 14:23:58 crc kubenswrapper[4824]: I1124 14:23:58.088507 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njkn6" event={"ID":"dface4f6-f5da-409b-879d-a6057f430e79","Type":"ContainerStarted","Data":"6c8825082b89e505290a52c678549c39fd43cd54e973ba41c745b42d2555c7da"} Nov 24 14:24:00 crc kubenswrapper[4824]: I1124 14:24:00.106351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njkn6" event={"ID":"dface4f6-f5da-409b-879d-a6057f430e79","Type":"ContainerStarted","Data":"d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50"} Nov 24 14:24:01 crc kubenswrapper[4824]: I1124 14:24:01.118257 4824 generic.go:334] "Generic (PLEG): container finished" podID="dface4f6-f5da-409b-879d-a6057f430e79" containerID="d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50" exitCode=0 Nov 24 14:24:01 crc kubenswrapper[4824]: I1124 14:24:01.118308 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njkn6" event={"ID":"dface4f6-f5da-409b-879d-a6057f430e79","Type":"ContainerDied","Data":"d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50"} Nov 24 14:24:03 crc kubenswrapper[4824]: I1124 14:24:03.141522 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njkn6" event={"ID":"dface4f6-f5da-409b-879d-a6057f430e79","Type":"ContainerStarted","Data":"b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95"} Nov 24 14:24:03 crc kubenswrapper[4824]: I1124 14:24:03.171140 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-njkn6" podStartSLOduration=6.348160911 podStartE2EDuration="10.171119304s" podCreationTimestamp="2025-11-24 14:23:53 +0000 UTC" firstStartedPulling="2025-11-24 14:23:58.091019667 +0000 UTC m=+4179.730559007" lastFinishedPulling="2025-11-24 14:24:01.91397807 +0000 UTC m=+4183.553517400" observedRunningTime="2025-11-24 14:24:03.159084062 +0000 UTC m=+4184.798623382" watchObservedRunningTime="2025-11-24 14:24:03.171119304 +0000 UTC m=+4184.810658614" Nov 24 14:24:04 crc kubenswrapper[4824]: I1124 14:24:04.298237 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:24:04 crc kubenswrapper[4824]: I1124 14:24:04.298742 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:24:04 crc kubenswrapper[4824]: I1124 14:24:04.350267 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:24:14 crc kubenswrapper[4824]: I1124 14:24:14.352748 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:24:14 crc kubenswrapper[4824]: I1124 14:24:14.408881 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-njkn6"] Nov 24 14:24:15 crc kubenswrapper[4824]: I1124 14:24:15.244016 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-njkn6" podUID="dface4f6-f5da-409b-879d-a6057f430e79" containerName="registry-server" containerID="cri-o://b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95" gracePeriod=2 Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.213866 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.253446 4824 generic.go:334] "Generic (PLEG): container finished" podID="dface4f6-f5da-409b-879d-a6057f430e79" containerID="b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95" exitCode=0 Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.253490 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njkn6" event={"ID":"dface4f6-f5da-409b-879d-a6057f430e79","Type":"ContainerDied","Data":"b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95"} Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.253500 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njkn6" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.253519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njkn6" event={"ID":"dface4f6-f5da-409b-879d-a6057f430e79","Type":"ContainerDied","Data":"6c8825082b89e505290a52c678549c39fd43cd54e973ba41c745b42d2555c7da"} Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.253539 4824 scope.go:117] "RemoveContainer" containerID="b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.275970 4824 scope.go:117] "RemoveContainer" containerID="d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.304001 4824 scope.go:117] "RemoveContainer" containerID="f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.310087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-catalog-content\") pod \"dface4f6-f5da-409b-879d-a6057f430e79\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.310234 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddchj\" (UniqueName: \"kubernetes.io/projected/dface4f6-f5da-409b-879d-a6057f430e79-kube-api-access-ddchj\") pod \"dface4f6-f5da-409b-879d-a6057f430e79\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.310251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-utilities\") pod \"dface4f6-f5da-409b-879d-a6057f430e79\" (UID: \"dface4f6-f5da-409b-879d-a6057f430e79\") " Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.311582 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-utilities" (OuterVolumeSpecName: "utilities") pod "dface4f6-f5da-409b-879d-a6057f430e79" (UID: "dface4f6-f5da-409b-879d-a6057f430e79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.317042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dface4f6-f5da-409b-879d-a6057f430e79-kube-api-access-ddchj" (OuterVolumeSpecName: "kube-api-access-ddchj") pod "dface4f6-f5da-409b-879d-a6057f430e79" (UID: "dface4f6-f5da-409b-879d-a6057f430e79"). InnerVolumeSpecName "kube-api-access-ddchj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.334331 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dface4f6-f5da-409b-879d-a6057f430e79" (UID: "dface4f6-f5da-409b-879d-a6057f430e79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.389285 4824 scope.go:117] "RemoveContainer" containerID="b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95" Nov 24 14:24:16 crc kubenswrapper[4824]: E1124 14:24:16.389818 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95\": container with ID starting with b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95 not found: ID does not exist" containerID="b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.389898 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95"} err="failed to get container status \"b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95\": rpc error: code = NotFound desc = could not find container \"b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95\": container with ID starting with b94f690a1a650def7a0656c7ca31d01c5ccc24ebbbdb5a299eec5b5b30cdad95 not found: ID does not exist" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.390003 4824 scope.go:117] "RemoveContainer" containerID="d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50" Nov 24 14:24:16 crc kubenswrapper[4824]: E1124 14:24:16.390431 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50\": container with ID starting with d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50 not found: ID does not exist" containerID="d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.390520 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50"} err="failed to get container status \"d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50\": rpc error: code = NotFound desc = could not find container \"d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50\": container with ID starting with d364407e89098c8f3c07f7e6f9e946ed93a68c1851322dda1d66858f2b3ecc50 not found: ID does not exist" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.390589 4824 scope.go:117] "RemoveContainer" containerID="f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5" Nov 24 14:24:16 crc kubenswrapper[4824]: E1124 14:24:16.390889 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5\": container with ID starting with f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5 not found: ID does not exist" containerID="f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.390967 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5"} err="failed to get container status \"f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5\": rpc error: code = NotFound desc = could not find container \"f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5\": container with ID starting with f07009d169821faad6cf641b07b9d269bc5179de730d4a48adc2bb32aba9c1c5 not found: ID does not exist" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.413598 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.413718 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddchj\" (UniqueName: \"kubernetes.io/projected/dface4f6-f5da-409b-879d-a6057f430e79-kube-api-access-ddchj\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.413777 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dface4f6-f5da-409b-879d-a6057f430e79-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.587736 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-njkn6"] Nov 24 14:24:16 crc kubenswrapper[4824]: I1124 14:24:16.596420 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-njkn6"] Nov 24 14:24:17 crc kubenswrapper[4824]: I1124 14:24:17.022551 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dface4f6-f5da-409b-879d-a6057f430e79" path="/var/lib/kubelet/pods/dface4f6-f5da-409b-879d-a6057f430e79/volumes" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.071178 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w2zbx"] Nov 24 14:24:38 crc kubenswrapper[4824]: E1124 14:24:38.072075 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dface4f6-f5da-409b-879d-a6057f430e79" containerName="registry-server" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.072087 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dface4f6-f5da-409b-879d-a6057f430e79" containerName="registry-server" Nov 24 14:24:38 crc kubenswrapper[4824]: E1124 14:24:38.072116 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dface4f6-f5da-409b-879d-a6057f430e79" containerName="extract-utilities" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.072123 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dface4f6-f5da-409b-879d-a6057f430e79" containerName="extract-utilities" Nov 24 14:24:38 crc kubenswrapper[4824]: E1124 14:24:38.072138 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dface4f6-f5da-409b-879d-a6057f430e79" containerName="extract-content" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.072143 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dface4f6-f5da-409b-879d-a6057f430e79" containerName="extract-content" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.072327 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="dface4f6-f5da-409b-879d-a6057f430e79" containerName="registry-server" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.073614 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.096073 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w2zbx"] Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.147372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-catalog-content\") pod \"community-operators-w2zbx\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.147449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22cf6\" (UniqueName: \"kubernetes.io/projected/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-kube-api-access-22cf6\") pod \"community-operators-w2zbx\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.147581 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-utilities\") pod \"community-operators-w2zbx\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.249643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-utilities\") pod \"community-operators-w2zbx\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.249758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-catalog-content\") pod \"community-operators-w2zbx\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.249833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22cf6\" (UniqueName: \"kubernetes.io/projected/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-kube-api-access-22cf6\") pod \"community-operators-w2zbx\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.250244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-utilities\") pod \"community-operators-w2zbx\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.250298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-catalog-content\") pod \"community-operators-w2zbx\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.278512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22cf6\" (UniqueName: \"kubernetes.io/projected/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-kube-api-access-22cf6\") pod \"community-operators-w2zbx\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:38 crc kubenswrapper[4824]: I1124 14:24:38.401525 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:39 crc kubenswrapper[4824]: I1124 14:24:39.112891 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w2zbx"] Nov 24 14:24:39 crc kubenswrapper[4824]: I1124 14:24:39.516350 4824 generic.go:334] "Generic (PLEG): container finished" podID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerID="157d31c34a940893a5a6b54c2ed3539dfee5fc91be725ae1fa24c6a6c0c73eaf" exitCode=0 Nov 24 14:24:39 crc kubenswrapper[4824]: I1124 14:24:39.516415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zbx" event={"ID":"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce","Type":"ContainerDied","Data":"157d31c34a940893a5a6b54c2ed3539dfee5fc91be725ae1fa24c6a6c0c73eaf"} Nov 24 14:24:39 crc kubenswrapper[4824]: I1124 14:24:39.516452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zbx" event={"ID":"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce","Type":"ContainerStarted","Data":"708b0cd69c288493ce737d4f0dadda48ed897247e71e557800325a8f98b19dbc"} Nov 24 14:24:41 crc kubenswrapper[4824]: I1124 14:24:41.536529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zbx" event={"ID":"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce","Type":"ContainerStarted","Data":"970d14fd5a58313d9f9d9d41cb9ec923408d72480c0a1e77d2811195107517a5"} Nov 24 14:24:45 crc kubenswrapper[4824]: I1124 14:24:45.575453 4824 generic.go:334] "Generic (PLEG): container finished" podID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerID="970d14fd5a58313d9f9d9d41cb9ec923408d72480c0a1e77d2811195107517a5" exitCode=0 Nov 24 14:24:45 crc kubenswrapper[4824]: I1124 14:24:45.575556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zbx" event={"ID":"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce","Type":"ContainerDied","Data":"970d14fd5a58313d9f9d9d41cb9ec923408d72480c0a1e77d2811195107517a5"} Nov 24 14:24:45 crc kubenswrapper[4824]: I1124 14:24:45.579412 4824 generic.go:334] "Generic (PLEG): container finished" podID="7cd566d4-13a3-423a-894d-d8dcb52325f7" containerID="278a250b83609700c67e370907958cbc5223716a9a61a77f196c35a24fb14670" exitCode=0 Nov 24 14:24:45 crc kubenswrapper[4824]: I1124 14:24:45.579482 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/crc-debug-gx82j" event={"ID":"7cd566d4-13a3-423a-894d-d8dcb52325f7","Type":"ContainerDied","Data":"278a250b83609700c67e370907958cbc5223716a9a61a77f196c35a24fb14670"} Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.589267 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zbx" event={"ID":"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce","Type":"ContainerStarted","Data":"0c7a150b5b7a2e0fbb8ef97cca1720d6520432d701c94893263318b6b738bad6"} Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.623638 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w2zbx" podStartSLOduration=1.99014719 podStartE2EDuration="8.623614441s" podCreationTimestamp="2025-11-24 14:24:38 +0000 UTC" firstStartedPulling="2025-11-24 14:24:39.521251177 +0000 UTC m=+4221.160790477" lastFinishedPulling="2025-11-24 14:24:46.154718418 +0000 UTC m=+4227.794257728" observedRunningTime="2025-11-24 14:24:46.611916898 +0000 UTC m=+4228.251456208" watchObservedRunningTime="2025-11-24 14:24:46.623614441 +0000 UTC m=+4228.263153751" Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.709585 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.753445 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hk622/crc-debug-gx82j"] Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.764391 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hk622/crc-debug-gx82j"] Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.812218 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzk92\" (UniqueName: \"kubernetes.io/projected/7cd566d4-13a3-423a-894d-d8dcb52325f7-kube-api-access-pzk92\") pod \"7cd566d4-13a3-423a-894d-d8dcb52325f7\" (UID: \"7cd566d4-13a3-423a-894d-d8dcb52325f7\") " Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.812571 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cd566d4-13a3-423a-894d-d8dcb52325f7-host\") pod \"7cd566d4-13a3-423a-894d-d8dcb52325f7\" (UID: \"7cd566d4-13a3-423a-894d-d8dcb52325f7\") " Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.812704 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cd566d4-13a3-423a-894d-d8dcb52325f7-host" (OuterVolumeSpecName: "host") pod "7cd566d4-13a3-423a-894d-d8dcb52325f7" (UID: "7cd566d4-13a3-423a-894d-d8dcb52325f7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.813367 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cd566d4-13a3-423a-894d-d8dcb52325f7-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.817680 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cd566d4-13a3-423a-894d-d8dcb52325f7-kube-api-access-pzk92" (OuterVolumeSpecName: "kube-api-access-pzk92") pod "7cd566d4-13a3-423a-894d-d8dcb52325f7" (UID: "7cd566d4-13a3-423a-894d-d8dcb52325f7"). InnerVolumeSpecName "kube-api-access-pzk92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:24:46 crc kubenswrapper[4824]: I1124 14:24:46.914634 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzk92\" (UniqueName: \"kubernetes.io/projected/7cd566d4-13a3-423a-894d-d8dcb52325f7-kube-api-access-pzk92\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:47 crc kubenswrapper[4824]: I1124 14:24:47.021972 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cd566d4-13a3-423a-894d-d8dcb52325f7" path="/var/lib/kubelet/pods/7cd566d4-13a3-423a-894d-d8dcb52325f7/volumes" Nov 24 14:24:47 crc kubenswrapper[4824]: I1124 14:24:47.599403 4824 scope.go:117] "RemoveContainer" containerID="278a250b83609700c67e370907958cbc5223716a9a61a77f196c35a24fb14670" Nov 24 14:24:47 crc kubenswrapper[4824]: I1124 14:24:47.599426 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-gx82j" Nov 24 14:24:47 crc kubenswrapper[4824]: I1124 14:24:47.995012 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hk622/crc-debug-vpbzk"] Nov 24 14:24:47 crc kubenswrapper[4824]: E1124 14:24:47.995777 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cd566d4-13a3-423a-894d-d8dcb52325f7" containerName="container-00" Nov 24 14:24:47 crc kubenswrapper[4824]: I1124 14:24:47.995803 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cd566d4-13a3-423a-894d-d8dcb52325f7" containerName="container-00" Nov 24 14:24:47 crc kubenswrapper[4824]: I1124 14:24:47.996074 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cd566d4-13a3-423a-894d-d8dcb52325f7" containerName="container-00" Nov 24 14:24:47 crc kubenswrapper[4824]: I1124 14:24:47.996779 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.039702 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-host\") pod \"crc-debug-vpbzk\" (UID: \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\") " pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.039761 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvtjn\" (UniqueName: \"kubernetes.io/projected/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-kube-api-access-nvtjn\") pod \"crc-debug-vpbzk\" (UID: \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\") " pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.141186 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-host\") pod \"crc-debug-vpbzk\" (UID: \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\") " pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.141256 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvtjn\" (UniqueName: \"kubernetes.io/projected/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-kube-api-access-nvtjn\") pod \"crc-debug-vpbzk\" (UID: \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\") " pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.141374 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-host\") pod \"crc-debug-vpbzk\" (UID: \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\") " pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.164608 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvtjn\" (UniqueName: \"kubernetes.io/projected/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-kube-api-access-nvtjn\") pod \"crc-debug-vpbzk\" (UID: \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\") " pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.316873 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:48 crc kubenswrapper[4824]: W1124 14:24:48.354776 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e0a4e12_ce46_4b07_86d7_b4f15804c40d.slice/crio-8b257bb01bcb8b6af418549e56c5f0e2b4fe2c019c66a522a2b5d2f72aa3a6b7 WatchSource:0}: Error finding container 8b257bb01bcb8b6af418549e56c5f0e2b4fe2c019c66a522a2b5d2f72aa3a6b7: Status 404 returned error can't find the container with id 8b257bb01bcb8b6af418549e56c5f0e2b4fe2c019c66a522a2b5d2f72aa3a6b7 Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.402664 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.405695 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.464267 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.622627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/crc-debug-vpbzk" event={"ID":"8e0a4e12-ce46-4b07-86d7-b4f15804c40d","Type":"ContainerStarted","Data":"396e5d811bfd763d037363bcb502dd5355911ad5874fc306936b32734a95fb50"} Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.622941 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/crc-debug-vpbzk" event={"ID":"8e0a4e12-ce46-4b07-86d7-b4f15804c40d","Type":"ContainerStarted","Data":"8b257bb01bcb8b6af418549e56c5f0e2b4fe2c019c66a522a2b5d2f72aa3a6b7"} Nov 24 14:24:48 crc kubenswrapper[4824]: I1124 14:24:48.645520 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hk622/crc-debug-vpbzk" podStartSLOduration=1.645488758 podStartE2EDuration="1.645488758s" podCreationTimestamp="2025-11-24 14:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:24:48.640622406 +0000 UTC m=+4230.280161726" watchObservedRunningTime="2025-11-24 14:24:48.645488758 +0000 UTC m=+4230.285028078" Nov 24 14:24:49 crc kubenswrapper[4824]: I1124 14:24:49.629646 4824 generic.go:334] "Generic (PLEG): container finished" podID="8e0a4e12-ce46-4b07-86d7-b4f15804c40d" containerID="396e5d811bfd763d037363bcb502dd5355911ad5874fc306936b32734a95fb50" exitCode=0 Nov 24 14:24:49 crc kubenswrapper[4824]: I1124 14:24:49.629729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/crc-debug-vpbzk" event={"ID":"8e0a4e12-ce46-4b07-86d7-b4f15804c40d","Type":"ContainerDied","Data":"396e5d811bfd763d037363bcb502dd5355911ad5874fc306936b32734a95fb50"} Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.169283 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.199999 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hk622/crc-debug-vpbzk"] Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.207126 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hk622/crc-debug-vpbzk"] Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.299589 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-host\") pod \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\" (UID: \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\") " Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.299732 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-host" (OuterVolumeSpecName: "host") pod "8e0a4e12-ce46-4b07-86d7-b4f15804c40d" (UID: "8e0a4e12-ce46-4b07-86d7-b4f15804c40d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.299745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvtjn\" (UniqueName: \"kubernetes.io/projected/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-kube-api-access-nvtjn\") pod \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\" (UID: \"8e0a4e12-ce46-4b07-86d7-b4f15804c40d\") " Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.300639 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.304763 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-kube-api-access-nvtjn" (OuterVolumeSpecName: "kube-api-access-nvtjn") pod "8e0a4e12-ce46-4b07-86d7-b4f15804c40d" (UID: "8e0a4e12-ce46-4b07-86d7-b4f15804c40d"). InnerVolumeSpecName "kube-api-access-nvtjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.402530 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvtjn\" (UniqueName: \"kubernetes.io/projected/8e0a4e12-ce46-4b07-86d7-b4f15804c40d-kube-api-access-nvtjn\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.645797 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b257bb01bcb8b6af418549e56c5f0e2b4fe2c019c66a522a2b5d2f72aa3a6b7" Nov 24 14:24:51 crc kubenswrapper[4824]: I1124 14:24:51.645878 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-vpbzk" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.424695 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hk622/crc-debug-zv74b"] Nov 24 14:24:52 crc kubenswrapper[4824]: E1124 14:24:52.425540 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0a4e12-ce46-4b07-86d7-b4f15804c40d" containerName="container-00" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.425559 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0a4e12-ce46-4b07-86d7-b4f15804c40d" containerName="container-00" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.425838 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e0a4e12-ce46-4b07-86d7-b4f15804c40d" containerName="container-00" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.426874 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.522218 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d19e33b-fb26-4e6b-851d-9fd333710c5d-host\") pod \"crc-debug-zv74b\" (UID: \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\") " pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.522455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdgfx\" (UniqueName: \"kubernetes.io/projected/0d19e33b-fb26-4e6b-851d-9fd333710c5d-kube-api-access-hdgfx\") pod \"crc-debug-zv74b\" (UID: \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\") " pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.624490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdgfx\" (UniqueName: \"kubernetes.io/projected/0d19e33b-fb26-4e6b-851d-9fd333710c5d-kube-api-access-hdgfx\") pod \"crc-debug-zv74b\" (UID: \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\") " pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.624589 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d19e33b-fb26-4e6b-851d-9fd333710c5d-host\") pod \"crc-debug-zv74b\" (UID: \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\") " pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.624765 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d19e33b-fb26-4e6b-851d-9fd333710c5d-host\") pod \"crc-debug-zv74b\" (UID: \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\") " pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.644052 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdgfx\" (UniqueName: \"kubernetes.io/projected/0d19e33b-fb26-4e6b-851d-9fd333710c5d-kube-api-access-hdgfx\") pod \"crc-debug-zv74b\" (UID: \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\") " pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:52 crc kubenswrapper[4824]: I1124 14:24:52.746587 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:52 crc kubenswrapper[4824]: W1124 14:24:52.789734 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d19e33b_fb26_4e6b_851d_9fd333710c5d.slice/crio-4d6281af9d172475a380049932006a31fc8a7c0c8c0c37ad13110b1918c06a8f WatchSource:0}: Error finding container 4d6281af9d172475a380049932006a31fc8a7c0c8c0c37ad13110b1918c06a8f: Status 404 returned error can't find the container with id 4d6281af9d172475a380049932006a31fc8a7c0c8c0c37ad13110b1918c06a8f Nov 24 14:24:53 crc kubenswrapper[4824]: I1124 14:24:53.020759 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e0a4e12-ce46-4b07-86d7-b4f15804c40d" path="/var/lib/kubelet/pods/8e0a4e12-ce46-4b07-86d7-b4f15804c40d/volumes" Nov 24 14:24:53 crc kubenswrapper[4824]: I1124 14:24:53.672745 4824 generic.go:334] "Generic (PLEG): container finished" podID="0d19e33b-fb26-4e6b-851d-9fd333710c5d" containerID="9e96fdfe1bc9449a0571860865c571515ac540383cf084bf5c4c2ac38f5adfd5" exitCode=0 Nov 24 14:24:53 crc kubenswrapper[4824]: I1124 14:24:53.672793 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/crc-debug-zv74b" event={"ID":"0d19e33b-fb26-4e6b-851d-9fd333710c5d","Type":"ContainerDied","Data":"9e96fdfe1bc9449a0571860865c571515ac540383cf084bf5c4c2ac38f5adfd5"} Nov 24 14:24:53 crc kubenswrapper[4824]: I1124 14:24:53.672870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/crc-debug-zv74b" event={"ID":"0d19e33b-fb26-4e6b-851d-9fd333710c5d","Type":"ContainerStarted","Data":"4d6281af9d172475a380049932006a31fc8a7c0c8c0c37ad13110b1918c06a8f"} Nov 24 14:24:53 crc kubenswrapper[4824]: I1124 14:24:53.718917 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hk622/crc-debug-zv74b"] Nov 24 14:24:53 crc kubenswrapper[4824]: I1124 14:24:53.729804 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hk622/crc-debug-zv74b"] Nov 24 14:24:54 crc kubenswrapper[4824]: I1124 14:24:54.786487 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:54 crc kubenswrapper[4824]: I1124 14:24:54.969889 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d19e33b-fb26-4e6b-851d-9fd333710c5d-host\") pod \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\" (UID: \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\") " Nov 24 14:24:54 crc kubenswrapper[4824]: I1124 14:24:54.970138 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdgfx\" (UniqueName: \"kubernetes.io/projected/0d19e33b-fb26-4e6b-851d-9fd333710c5d-kube-api-access-hdgfx\") pod \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\" (UID: \"0d19e33b-fb26-4e6b-851d-9fd333710c5d\") " Nov 24 14:24:54 crc kubenswrapper[4824]: I1124 14:24:54.969985 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d19e33b-fb26-4e6b-851d-9fd333710c5d-host" (OuterVolumeSpecName: "host") pod "0d19e33b-fb26-4e6b-851d-9fd333710c5d" (UID: "0d19e33b-fb26-4e6b-851d-9fd333710c5d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:24:54 crc kubenswrapper[4824]: I1124 14:24:54.970643 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d19e33b-fb26-4e6b-851d-9fd333710c5d-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:54 crc kubenswrapper[4824]: I1124 14:24:54.978039 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d19e33b-fb26-4e6b-851d-9fd333710c5d-kube-api-access-hdgfx" (OuterVolumeSpecName: "kube-api-access-hdgfx") pod "0d19e33b-fb26-4e6b-851d-9fd333710c5d" (UID: "0d19e33b-fb26-4e6b-851d-9fd333710c5d"). InnerVolumeSpecName "kube-api-access-hdgfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:24:55 crc kubenswrapper[4824]: I1124 14:24:55.021249 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d19e33b-fb26-4e6b-851d-9fd333710c5d" path="/var/lib/kubelet/pods/0d19e33b-fb26-4e6b-851d-9fd333710c5d/volumes" Nov 24 14:24:55 crc kubenswrapper[4824]: I1124 14:24:55.072609 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdgfx\" (UniqueName: \"kubernetes.io/projected/0d19e33b-fb26-4e6b-851d-9fd333710c5d-kube-api-access-hdgfx\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:55 crc kubenswrapper[4824]: I1124 14:24:55.693559 4824 scope.go:117] "RemoveContainer" containerID="9e96fdfe1bc9449a0571860865c571515ac540383cf084bf5c4c2ac38f5adfd5" Nov 24 14:24:55 crc kubenswrapper[4824]: I1124 14:24:55.693712 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/crc-debug-zv74b" Nov 24 14:24:58 crc kubenswrapper[4824]: I1124 14:24:58.461479 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:24:58 crc kubenswrapper[4824]: I1124 14:24:58.511636 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w2zbx"] Nov 24 14:24:58 crc kubenswrapper[4824]: I1124 14:24:58.717101 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w2zbx" podUID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerName="registry-server" containerID="cri-o://0c7a150b5b7a2e0fbb8ef97cca1720d6520432d701c94893263318b6b738bad6" gracePeriod=2 Nov 24 14:24:59 crc kubenswrapper[4824]: I1124 14:24:59.726672 4824 generic.go:334] "Generic (PLEG): container finished" podID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerID="0c7a150b5b7a2e0fbb8ef97cca1720d6520432d701c94893263318b6b738bad6" exitCode=0 Nov 24 14:24:59 crc kubenswrapper[4824]: I1124 14:24:59.726752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zbx" event={"ID":"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce","Type":"ContainerDied","Data":"0c7a150b5b7a2e0fbb8ef97cca1720d6520432d701c94893263318b6b738bad6"} Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.224299 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.371075 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-catalog-content\") pod \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.371148 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22cf6\" (UniqueName: \"kubernetes.io/projected/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-kube-api-access-22cf6\") pod \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.371176 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-utilities\") pod \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\" (UID: \"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce\") " Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.374181 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-utilities" (OuterVolumeSpecName: "utilities") pod "065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" (UID: "065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.378348 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-kube-api-access-22cf6" (OuterVolumeSpecName: "kube-api-access-22cf6") pod "065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" (UID: "065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce"). InnerVolumeSpecName "kube-api-access-22cf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.455299 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" (UID: "065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.473178 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.473221 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22cf6\" (UniqueName: \"kubernetes.io/projected/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-kube-api-access-22cf6\") on node \"crc\" DevicePath \"\"" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.473235 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.736678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zbx" event={"ID":"065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce","Type":"ContainerDied","Data":"708b0cd69c288493ce737d4f0dadda48ed897247e71e557800325a8f98b19dbc"} Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.736729 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2zbx" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.736747 4824 scope.go:117] "RemoveContainer" containerID="0c7a150b5b7a2e0fbb8ef97cca1720d6520432d701c94893263318b6b738bad6" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.754166 4824 scope.go:117] "RemoveContainer" containerID="970d14fd5a58313d9f9d9d41cb9ec923408d72480c0a1e77d2811195107517a5" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.778922 4824 scope.go:117] "RemoveContainer" containerID="157d31c34a940893a5a6b54c2ed3539dfee5fc91be725ae1fa24c6a6c0c73eaf" Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.800399 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w2zbx"] Nov 24 14:25:00 crc kubenswrapper[4824]: I1124 14:25:00.812030 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w2zbx"] Nov 24 14:25:01 crc kubenswrapper[4824]: I1124 14:25:01.021459 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" path="/var/lib/kubelet/pods/065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce/volumes" Nov 24 14:25:13 crc kubenswrapper[4824]: I1124 14:25:13.057027 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7545588bcd-5szvs_d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960/barbican-api/0.log" Nov 24 14:25:13 crc kubenswrapper[4824]: I1124 14:25:13.267822 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7545588bcd-5szvs_d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960/barbican-api-log/0.log" Nov 24 14:25:13 crc kubenswrapper[4824]: I1124 14:25:13.357391 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c785d8464-nfqng_6534eeea-5982-4010-a073-f307d7d58036/barbican-keystone-listener/0.log" Nov 24 14:25:13 crc kubenswrapper[4824]: I1124 14:25:13.441589 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c785d8464-nfqng_6534eeea-5982-4010-a073-f307d7d58036/barbican-keystone-listener-log/0.log" Nov 24 14:25:13 crc kubenswrapper[4824]: I1124 14:25:13.605627 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b664bf49f-czcvh_b227cafd-bcc4-4e74-96ea-0935efd47fcc/barbican-worker-log/0.log" Nov 24 14:25:13 crc kubenswrapper[4824]: I1124 14:25:13.617545 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b664bf49f-czcvh_b227cafd-bcc4-4e74-96ea-0935efd47fcc/barbican-worker/0.log" Nov 24 14:25:13 crc kubenswrapper[4824]: I1124 14:25:13.797981 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs_5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:13 crc kubenswrapper[4824]: I1124 14:25:13.848169 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f521f794-dddf-4c12-9b64-665b541d43ec/ceilometer-central-agent/0.log" Nov 24 14:25:13 crc kubenswrapper[4824]: I1124 14:25:13.976428 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f521f794-dddf-4c12-9b64-665b541d43ec/ceilometer-notification-agent/0.log" Nov 24 14:25:14 crc kubenswrapper[4824]: I1124 14:25:14.038284 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f521f794-dddf-4c12-9b64-665b541d43ec/sg-core/0.log" Nov 24 14:25:14 crc kubenswrapper[4824]: I1124 14:25:14.045909 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f521f794-dddf-4c12-9b64-665b541d43ec/proxy-httpd/0.log" Nov 24 14:25:14 crc kubenswrapper[4824]: I1124 14:25:14.223260 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4e42a7c2-0db7-4c5d-b731-b543c875ebb0/cinder-api/0.log" Nov 24 14:25:14 crc kubenswrapper[4824]: I1124 14:25:14.238930 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4e42a7c2-0db7-4c5d-b731-b543c875ebb0/cinder-api-log/0.log" Nov 24 14:25:14 crc kubenswrapper[4824]: I1124 14:25:14.374499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_51aacb45-4a5f-42d3-b905-a051a14856f4/cinder-scheduler/0.log" Nov 24 14:25:14 crc kubenswrapper[4824]: I1124 14:25:14.487421 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_51aacb45-4a5f-42d3-b905-a051a14856f4/probe/0.log" Nov 24 14:25:15 crc kubenswrapper[4824]: I1124 14:25:15.193069 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-k5rds_18493374-ba71-4ac2-97dd-171111137b81/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:15 crc kubenswrapper[4824]: I1124 14:25:15.320974 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8_92ed5101-5b03-4be3-bd82-32ba9915ba15/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:15 crc kubenswrapper[4824]: I1124 14:25:15.496685 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-hc64g_0bec1542-f209-4da4-a540-9be4708ab637/init/0.log" Nov 24 14:25:15 crc kubenswrapper[4824]: I1124 14:25:15.704430 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-hc64g_0bec1542-f209-4da4-a540-9be4708ab637/init/0.log" Nov 24 14:25:15 crc kubenswrapper[4824]: I1124 14:25:15.812085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm_73fc4178-cf60-40cb-b3f0-ea7dc1b1bada/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:16 crc kubenswrapper[4824]: I1124 14:25:16.006633 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-hc64g_0bec1542-f209-4da4-a540-9be4708ab637/dnsmasq-dns/0.log" Nov 24 14:25:16 crc kubenswrapper[4824]: I1124 14:25:16.162830 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f7321bd0-b6b4-45dc-a4c6-c7a26b601b17/glance-httpd/0.log" Nov 24 14:25:16 crc kubenswrapper[4824]: I1124 14:25:16.174764 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f7321bd0-b6b4-45dc-a4c6-c7a26b601b17/glance-log/0.log" Nov 24 14:25:16 crc kubenswrapper[4824]: I1124 14:25:16.332350 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_dda48846-9717-4dd4-b4af-afe6887923bb/glance-httpd/0.log" Nov 24 14:25:17 crc kubenswrapper[4824]: I1124 14:25:17.074601 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_dda48846-9717-4dd4-b4af-afe6887923bb/glance-log/0.log" Nov 24 14:25:17 crc kubenswrapper[4824]: I1124 14:25:17.172328 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-959bb6dd8-8pcrj_5828a51f-d146-44ea-b8ea-423e017939d8/horizon/1.log" Nov 24 14:25:17 crc kubenswrapper[4824]: I1124 14:25:17.299361 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-959bb6dd8-8pcrj_5828a51f-d146-44ea-b8ea-423e017939d8/horizon/0.log" Nov 24 14:25:17 crc kubenswrapper[4824]: I1124 14:25:17.570338 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv_27890aad-2f36-400e-8535-afecafb22a37/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:17 crc kubenswrapper[4824]: I1124 14:25:17.669205 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nhzbn_27ecdc8a-4391-4eb5-a900-0f28984fd1d4/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:17 crc kubenswrapper[4824]: I1124 14:25:17.673064 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-959bb6dd8-8pcrj_5828a51f-d146-44ea-b8ea-423e017939d8/horizon-log/0.log" Nov 24 14:25:18 crc kubenswrapper[4824]: I1124 14:25:18.195536 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399881-7g6ln_d24ebeb5-8ce9-42cb-95c6-29162b65eb35/keystone-cron/0.log" Nov 24 14:25:18 crc kubenswrapper[4824]: I1124 14:25:18.306330 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-57d8799bcb-x76nj_378c23b7-2f70-47a0-b605-67087ed59e70/keystone-api/0.log" Nov 24 14:25:18 crc kubenswrapper[4824]: I1124 14:25:18.361706 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8a804c88-d42d-43ab-b316-d196ba26f5d1/kube-state-metrics/0.log" Nov 24 14:25:18 crc kubenswrapper[4824]: I1124 14:25:18.607273 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-smm76_fa9782de-a426-4c52-a142-335c78e55a33/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:18 crc kubenswrapper[4824]: I1124 14:25:18.998892 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt_18802c46-869d-4e00-848e-79df8665851d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:19 crc kubenswrapper[4824]: I1124 14:25:19.335571 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-798f5d656f-r9xcp_bf77fc79-5c9e-4688-be73-b390bef539fd/neutron-httpd/0.log" Nov 24 14:25:19 crc kubenswrapper[4824]: I1124 14:25:19.489667 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-798f5d656f-r9xcp_bf77fc79-5c9e-4688-be73-b390bef539fd/neutron-api/0.log" Nov 24 14:25:20 crc kubenswrapper[4824]: I1124 14:25:20.197634 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e5819ad8-0aa5-4d68-831e-09c741267fd5/nova-cell0-conductor-conductor/0.log" Nov 24 14:25:20 crc kubenswrapper[4824]: I1124 14:25:20.322989 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_73aeee43-4360-4af8-9c1e-e1d9bc712157/nova-cell1-conductor-conductor/0.log" Nov 24 14:25:20 crc kubenswrapper[4824]: I1124 14:25:20.617664 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ebf76616-c332-4ed1-a188-99843d4692d3/nova-api-log/0.log" Nov 24 14:25:20 crc kubenswrapper[4824]: I1124 14:25:20.723246 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ebf76616-c332-4ed1-a188-99843d4692d3/nova-api-api/0.log" Nov 24 14:25:20 crc kubenswrapper[4824]: I1124 14:25:20.724099 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9be10433-b401-4853-ad5a-f32b7668fde2/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 14:25:21 crc kubenswrapper[4824]: I1124 14:25:21.043392 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-qtdx4_4c37785a-1f6f-44c9-8673-b7f9f465682a/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:21 crc kubenswrapper[4824]: I1124 14:25:21.110175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8cd73a38-f125-4e8d-b902-f8bc08c1122a/nova-metadata-log/0.log" Nov 24 14:25:21 crc kubenswrapper[4824]: I1124 14:25:21.857737 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_90325e3b-c1cb-4244-a81b-fa666652304b/mysql-bootstrap/0.log" Nov 24 14:25:22 crc kubenswrapper[4824]: I1124 14:25:22.033588 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_51c22644-2997-44ff-ab0d-13ce1a067aa2/nova-scheduler-scheduler/0.log" Nov 24 14:25:22 crc kubenswrapper[4824]: I1124 14:25:22.380938 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_90325e3b-c1cb-4244-a81b-fa666652304b/galera/0.log" Nov 24 14:25:22 crc kubenswrapper[4824]: I1124 14:25:22.387823 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_90325e3b-c1cb-4244-a81b-fa666652304b/mysql-bootstrap/0.log" Nov 24 14:25:22 crc kubenswrapper[4824]: I1124 14:25:22.689499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1e8da990-0e2a-47fd-b7d1-dec77870b40c/mysql-bootstrap/0.log" Nov 24 14:25:22 crc kubenswrapper[4824]: I1124 14:25:22.819737 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1e8da990-0e2a-47fd-b7d1-dec77870b40c/mysql-bootstrap/0.log" Nov 24 14:25:22 crc kubenswrapper[4824]: I1124 14:25:22.868474 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1e8da990-0e2a-47fd-b7d1-dec77870b40c/galera/0.log" Nov 24 14:25:22 crc kubenswrapper[4824]: I1124 14:25:22.946277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8cd73a38-f125-4e8d-b902-f8bc08c1122a/nova-metadata-metadata/0.log" Nov 24 14:25:23 crc kubenswrapper[4824]: I1124 14:25:23.101088 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_7d1c1940-0b8b-4b09-ae2a-76168c5a873b/openstackclient/0.log" Nov 24 14:25:23 crc kubenswrapper[4824]: I1124 14:25:23.332029 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-wb2pp_e586afb0-4f13-437a-b00f-ea7e1f6d82ad/openstack-network-exporter/0.log" Nov 24 14:25:23 crc kubenswrapper[4824]: I1124 14:25:23.385344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ds8fs_b3c188e2-9a89-47fe-8efe-1187a3c81552/ovn-controller/0.log" Nov 24 14:25:23 crc kubenswrapper[4824]: I1124 14:25:23.592198 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jvdgl_25645cc4-c970-453f-a6c3-08545ce4c84c/ovsdb-server-init/0.log" Nov 24 14:25:23 crc kubenswrapper[4824]: I1124 14:25:23.810200 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jvdgl_25645cc4-c970-453f-a6c3-08545ce4c84c/ovsdb-server-init/0.log" Nov 24 14:25:23 crc kubenswrapper[4824]: I1124 14:25:23.862026 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jvdgl_25645cc4-c970-453f-a6c3-08545ce4c84c/ovsdb-server/0.log" Nov 24 14:25:23 crc kubenswrapper[4824]: I1124 14:25:23.919682 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jvdgl_25645cc4-c970-453f-a6c3-08545ce4c84c/ovs-vswitchd/0.log" Nov 24 14:25:24 crc kubenswrapper[4824]: I1124 14:25:24.635453 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_102b3dcf-b337-4f91-954b-a4996ba67ca7/ovn-northd/0.log" Nov 24 14:25:24 crc kubenswrapper[4824]: I1124 14:25:24.659397 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_102b3dcf-b337-4f91-954b-a4996ba67ca7/openstack-network-exporter/0.log" Nov 24 14:25:24 crc kubenswrapper[4824]: I1124 14:25:24.708633 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-bwlv7_f5eaa6e7-da63-4213-ac5d-57fac9f6cd63/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:24 crc kubenswrapper[4824]: I1124 14:25:24.904422 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7c045328-9e32-430c-b29b-91a3aeae62c7/openstack-network-exporter/0.log" Nov 24 14:25:24 crc kubenswrapper[4824]: I1124 14:25:24.980238 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7c045328-9e32-430c-b29b-91a3aeae62c7/ovsdbserver-nb/0.log" Nov 24 14:25:25 crc kubenswrapper[4824]: I1124 14:25:25.201515 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c11c797e-a7a6-4b6f-8261-cdb6bdc7a525/ovsdbserver-sb/0.log" Nov 24 14:25:25 crc kubenswrapper[4824]: I1124 14:25:25.201371 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c11c797e-a7a6-4b6f-8261-cdb6bdc7a525/openstack-network-exporter/0.log" Nov 24 14:25:25 crc kubenswrapper[4824]: I1124 14:25:25.805935 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bdb7b6678-pnfdp_d2b3f41b-4bd4-4fcf-b248-6efd3b68f512/placement-api/0.log" Nov 24 14:25:26 crc kubenswrapper[4824]: I1124 14:25:26.054129 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bdb7b6678-pnfdp_d2b3f41b-4bd4-4fcf-b248-6efd3b68f512/placement-log/0.log" Nov 24 14:25:26 crc kubenswrapper[4824]: I1124 14:25:26.070510 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_829719ca-f478-456b-9bbc-07cad8469731/setup-container/0.log" Nov 24 14:25:26 crc kubenswrapper[4824]: I1124 14:25:26.318719 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_829719ca-f478-456b-9bbc-07cad8469731/setup-container/0.log" Nov 24 14:25:26 crc kubenswrapper[4824]: I1124 14:25:26.344100 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_829719ca-f478-456b-9bbc-07cad8469731/rabbitmq/0.log" Nov 24 14:25:26 crc kubenswrapper[4824]: I1124 14:25:26.449543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98ec6c14-59af-4b38-8e01-89ff157036ca/setup-container/0.log" Nov 24 14:25:26 crc kubenswrapper[4824]: I1124 14:25:26.702847 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98ec6c14-59af-4b38-8e01-89ff157036ca/rabbitmq/0.log" Nov 24 14:25:26 crc kubenswrapper[4824]: I1124 14:25:26.775639 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h_20a6cb74-657c-42ba-a789-411e312d605b/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:26 crc kubenswrapper[4824]: I1124 14:25:26.832260 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98ec6c14-59af-4b38-8e01-89ff157036ca/setup-container/0.log" Nov 24 14:25:27 crc kubenswrapper[4824]: I1124 14:25:27.238679 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-94hhr_53c68f20-6a1f-449e-a54e-aa1df1bf7d32/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:27 crc kubenswrapper[4824]: I1124 14:25:27.300277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh_1a2dbe7f-7854-4dc7-9511-bba8a8c81a08/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:27 crc kubenswrapper[4824]: I1124 14:25:27.496226 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-97k6l_9ed4df3b-472c-4815-a5ef-66125b2dde81/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:27 crc kubenswrapper[4824]: I1124 14:25:27.633385 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-vtnl7_62b46948-d9cc-4a68-8d94-70a7538bd61d/ssh-known-hosts-edpm-deployment/0.log" Nov 24 14:25:27 crc kubenswrapper[4824]: I1124 14:25:27.939335 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8cc59fd67-2j94h_57b08eb1-c1bf-4b56-880f-4d8adeaa32fc/proxy-server/0.log" Nov 24 14:25:27 crc kubenswrapper[4824]: I1124 14:25:27.953311 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8cc59fd67-2j94h_57b08eb1-c1bf-4b56-880f-4d8adeaa32fc/proxy-httpd/0.log" Nov 24 14:25:28 crc kubenswrapper[4824]: I1124 14:25:28.107335 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-m5pdt_c2abd860-636d-4e37-8a85-334aefcf44ec/swift-ring-rebalance/0.log" Nov 24 14:25:28 crc kubenswrapper[4824]: I1124 14:25:28.300367 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/account-reaper/0.log" Nov 24 14:25:28 crc kubenswrapper[4824]: I1124 14:25:28.319035 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/account-auditor/0.log" Nov 24 14:25:28 crc kubenswrapper[4824]: I1124 14:25:28.472263 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/account-replicator/0.log" Nov 24 14:25:28 crc kubenswrapper[4824]: I1124 14:25:28.925755 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/account-server/0.log" Nov 24 14:25:28 crc kubenswrapper[4824]: I1124 14:25:28.996450 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/container-server/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.002208 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/container-replicator/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.038879 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/container-auditor/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.185483 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/container-updater/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.230639 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-auditor/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.288436 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-expirer/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.337529 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-replicator/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.579495 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-server/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.597599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/rsync/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.624559 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/swift-recon-cron/0.log" Nov 24 14:25:29 crc kubenswrapper[4824]: I1124 14:25:29.625411 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-updater/0.log" Nov 24 14:25:30 crc kubenswrapper[4824]: I1124 14:25:30.004648 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-k6x98_3fdb12da-712c-4ebe-97d1-4ebc46a1db80/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:30 crc kubenswrapper[4824]: I1124 14:25:30.123271 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0/tempest-tests-tempest-tests-runner/0.log" Nov 24 14:25:30 crc kubenswrapper[4824]: I1124 14:25:30.349298 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b719b5e7-33d8-43ec-9d61-19bd0676886a/test-operator-logs-container/0.log" Nov 24 14:25:30 crc kubenswrapper[4824]: I1124 14:25:30.464709 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s_ddb976c7-8261-4a6b-96a1-75a8e4b5d282/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:25:38 crc kubenswrapper[4824]: I1124 14:25:38.910135 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_a7687928-0bb1-456c-8555-e918db203197/memcached/0.log" Nov 24 14:26:00 crc kubenswrapper[4824]: I1124 14:26:00.281641 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-747cg_dc25960c-da62-4fe3-97c1-cf8526bc9c1d/kube-rbac-proxy/0.log" Nov 24 14:26:00 crc kubenswrapper[4824]: I1124 14:26:00.290668 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-747cg_dc25960c-da62-4fe3-97c1-cf8526bc9c1d/manager/0.log" Nov 24 14:26:01 crc kubenswrapper[4824]: I1124 14:26:01.050876 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-h5mx8_88cfde0a-21e8-45ca-a7b3-43699c2ff345/kube-rbac-proxy/0.log" Nov 24 14:26:01 crc kubenswrapper[4824]: I1124 14:26:01.086927 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-h5mx8_88cfde0a-21e8-45ca-a7b3-43699c2ff345/manager/0.log" Nov 24 14:26:01 crc kubenswrapper[4824]: I1124 14:26:01.275880 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-gw957_3b95ed18-ec44-4712-9f01-adeef55deb75/kube-rbac-proxy/0.log" Nov 24 14:26:01 crc kubenswrapper[4824]: I1124 14:26:01.336041 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-gw957_3b95ed18-ec44-4712-9f01-adeef55deb75/manager/0.log" Nov 24 14:26:01 crc kubenswrapper[4824]: I1124 14:26:01.459017 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/util/0.log" Nov 24 14:26:01 crc kubenswrapper[4824]: I1124 14:26:01.657232 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/util/0.log" Nov 24 14:26:01 crc kubenswrapper[4824]: I1124 14:26:01.675795 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/pull/0.log" Nov 24 14:26:01 crc kubenswrapper[4824]: I1124 14:26:01.680126 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/pull/0.log" Nov 24 14:26:01 crc kubenswrapper[4824]: I1124 14:26:01.845336 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/util/0.log" Nov 24 14:26:02 crc kubenswrapper[4824]: I1124 14:26:02.608398 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/pull/0.log" Nov 24 14:26:02 crc kubenswrapper[4824]: I1124 14:26:02.623184 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/extract/0.log" Nov 24 14:26:02 crc kubenswrapper[4824]: I1124 14:26:02.675905 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-gvhjd_9c973000-24b9-4087-a3eb-c5f087a50ae3/kube-rbac-proxy/0.log" Nov 24 14:26:02 crc kubenswrapper[4824]: I1124 14:26:02.882254 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-gvhjd_9c973000-24b9-4087-a3eb-c5f087a50ae3/manager/0.log" Nov 24 14:26:02 crc kubenswrapper[4824]: I1124 14:26:02.951164 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-s2k8r_0cac78b6-e909-4616-bce8-d3ce8ac90452/kube-rbac-proxy/0.log" Nov 24 14:26:02 crc kubenswrapper[4824]: I1124 14:26:02.984836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-s2k8r_0cac78b6-e909-4616-bce8-d3ce8ac90452/manager/0.log" Nov 24 14:26:03 crc kubenswrapper[4824]: I1124 14:26:03.166716 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-cdnsz_5f502855-0eb1-455d-8e3b-4af6ac353aa9/kube-rbac-proxy/0.log" Nov 24 14:26:03 crc kubenswrapper[4824]: I1124 14:26:03.266106 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-cdnsz_5f502855-0eb1-455d-8e3b-4af6ac353aa9/manager/0.log" Nov 24 14:26:03 crc kubenswrapper[4824]: I1124 14:26:03.333434 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-rn29q_ec73cd81-275a-44e0-93dc-6b20123a88d3/kube-rbac-proxy/0.log" Nov 24 14:26:03 crc kubenswrapper[4824]: I1124 14:26:03.572680 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-cmmdj_05a9eb12-a2a1-4fbe-901b-cb817e4713b4/kube-rbac-proxy/0.log" Nov 24 14:26:03 crc kubenswrapper[4824]: I1124 14:26:03.600531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-cmmdj_05a9eb12-a2a1-4fbe-901b-cb817e4713b4/manager/0.log" Nov 24 14:26:03 crc kubenswrapper[4824]: I1124 14:26:03.615875 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-rn29q_ec73cd81-275a-44e0-93dc-6b20123a88d3/manager/0.log" Nov 24 14:26:03 crc kubenswrapper[4824]: I1124 14:26:03.811956 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-v6dbs_65d269bf-f6af-4239-8e51-bf44fe7830d1/kube-rbac-proxy/0.log" Nov 24 14:26:03 crc kubenswrapper[4824]: I1124 14:26:03.911590 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-v6dbs_65d269bf-f6af-4239-8e51-bf44fe7830d1/manager/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.050347 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-xdrwd_3b3da19f-12b1-413d-b6aa-38bc498172c1/manager/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.106150 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-xdrwd_3b3da19f-12b1-413d-b6aa-38bc498172c1/kube-rbac-proxy/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.131782 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-mzs9m_9971e72a-bb18-4951-a752-ee96f1725801/kube-rbac-proxy/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.264308 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-mzs9m_9971e72a-bb18-4951-a752-ee96f1725801/manager/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.386419 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-zncj6_50437b59-29ab-4e49-a751-80a83d9e2aa0/kube-rbac-proxy/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.430147 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-zncj6_50437b59-29ab-4e49-a751-80a83d9e2aa0/manager/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.563654 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-qzxtx_4cca949a-2334-4d41-bb9e-13e6bd89d7a8/kube-rbac-proxy/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.640336 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-bg2m4_218ca72b-e49f-4830-8be8-9b6e85956009/kube-rbac-proxy/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.706507 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-qzxtx_4cca949a-2334-4d41-bb9e-13e6bd89d7a8/manager/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.721118 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-bg2m4_218ca72b-e49f-4830-8be8-9b6e85956009/manager/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.873851 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp_99379416-bd58-4014-a297-7b14b8ff4aeb/manager/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.894038 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp_99379416-bd58-4014-a297-7b14b8ff4aeb/kube-rbac-proxy/0.log" Nov 24 14:26:04 crc kubenswrapper[4824]: I1124 14:26:04.908390 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7547bd9b9-tpbcz_4f304aa2-86ae-45a3-9039-848548cbd500/kube-rbac-proxy/0.log" Nov 24 14:26:05 crc kubenswrapper[4824]: I1124 14:26:05.119043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-866b74d8b7-d68dq_af6dc481-60ca-48ae-a75c-edec34270b7f/kube-rbac-proxy/0.log" Nov 24 14:26:05 crc kubenswrapper[4824]: I1124 14:26:05.372199 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hht55_eaf96b33-49f7-4e4f-86e2-7a8a8b479bba/registry-server/0.log" Nov 24 14:26:05 crc kubenswrapper[4824]: I1124 14:26:05.375849 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-866b74d8b7-d68dq_af6dc481-60ca-48ae-a75c-edec34270b7f/operator/0.log" Nov 24 14:26:05 crc kubenswrapper[4824]: I1124 14:26:05.649362 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-2jjxl_926db806-d562-44f8-9ac9-71f337dbb202/kube-rbac-proxy/0.log" Nov 24 14:26:05 crc kubenswrapper[4824]: I1124 14:26:05.694540 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-2jjxl_926db806-d562-44f8-9ac9-71f337dbb202/manager/0.log" Nov 24 14:26:05 crc kubenswrapper[4824]: I1124 14:26:05.713836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-rzg64_4676e827-15a7-4286-a4a8-0147b70e315b/kube-rbac-proxy/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.010042 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-rzg64_4676e827-15a7-4286-a4a8-0147b70e315b/manager/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.097130 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-b82s2_c97af640-fa25-46a6-816a-104506790d6d/operator/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.118691 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7547bd9b9-tpbcz_4f304aa2-86ae-45a3-9039-848548cbd500/manager/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.157550 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-hnczv_39f22836-fab7-4ffa-b95f-d2dad15a98b9/kube-rbac-proxy/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.320238 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-hnczv_39f22836-fab7-4ffa-b95f-d2dad15a98b9/manager/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.328654 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9sh2g"] Nov 24 14:26:06 crc kubenswrapper[4824]: E1124 14:26:06.329066 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerName="extract-content" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.329084 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerName="extract-content" Nov 24 14:26:06 crc kubenswrapper[4824]: E1124 14:26:06.329098 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerName="registry-server" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.329104 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerName="registry-server" Nov 24 14:26:06 crc kubenswrapper[4824]: E1124 14:26:06.329119 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d19e33b-fb26-4e6b-851d-9fd333710c5d" containerName="container-00" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.329126 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d19e33b-fb26-4e6b-851d-9fd333710c5d" containerName="container-00" Nov 24 14:26:06 crc kubenswrapper[4824]: E1124 14:26:06.329143 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerName="extract-utilities" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.329149 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerName="extract-utilities" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.329308 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d19e33b-fb26-4e6b-851d-9fd333710c5d" containerName="container-00" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.329327 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="065d4aa3-c1fd-46ea-b3a2-dff3a5cd55ce" containerName="registry-server" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.330556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.339614 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-w82fg_3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2/kube-rbac-proxy/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.353223 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9sh2g"] Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.459562 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-w82fg_3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2/manager/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.494019 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-utilities\") pod \"certified-operators-9sh2g\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.494316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-catalog-content\") pod \"certified-operators-9sh2g\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.494580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85j92\" (UniqueName: \"kubernetes.io/projected/ff4d9122-d458-4f03-bb07-9c0a39833b2c-kube-api-access-85j92\") pod \"certified-operators-9sh2g\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.596541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-catalog-content\") pod \"certified-operators-9sh2g\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.596924 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85j92\" (UniqueName: \"kubernetes.io/projected/ff4d9122-d458-4f03-bb07-9c0a39833b2c-kube-api-access-85j92\") pod \"certified-operators-9sh2g\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.597000 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-utilities\") pod \"certified-operators-9sh2g\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.597378 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-catalog-content\") pod \"certified-operators-9sh2g\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.597480 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-utilities\") pod \"certified-operators-9sh2g\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.615158 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85j92\" (UniqueName: \"kubernetes.io/projected/ff4d9122-d458-4f03-bb07-9c0a39833b2c-kube-api-access-85j92\") pod \"certified-operators-9sh2g\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.648711 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.667774 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-fhn99_bc237c32-95f2-4386-88f3-a337d2269261/kube-rbac-proxy/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.669402 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-fhn99_bc237c32-95f2-4386-88f3-a337d2269261/manager/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.766322 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-t46st_c8001b29-70e7-4861-9b3d-41e732db7bf9/kube-rbac-proxy/0.log" Nov 24 14:26:06 crc kubenswrapper[4824]: I1124 14:26:06.897870 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-t46st_c8001b29-70e7-4861-9b3d-41e732db7bf9/manager/0.log" Nov 24 14:26:07 crc kubenswrapper[4824]: I1124 14:26:07.207598 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9sh2g"] Nov 24 14:26:07 crc kubenswrapper[4824]: I1124 14:26:07.392415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sh2g" event={"ID":"ff4d9122-d458-4f03-bb07-9c0a39833b2c","Type":"ContainerStarted","Data":"57d44192242257142864aa5bdaffc7c2e49bbdb8b05feff2bd7f65a51059b76c"} Nov 24 14:26:08 crc kubenswrapper[4824]: I1124 14:26:08.401509 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerID="a9808b1bc64d2af944468d2b1b376b35734f0513a7fc66c574c8fce225852f47" exitCode=0 Nov 24 14:26:08 crc kubenswrapper[4824]: I1124 14:26:08.401573 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sh2g" event={"ID":"ff4d9122-d458-4f03-bb07-9c0a39833b2c","Type":"ContainerDied","Data":"a9808b1bc64d2af944468d2b1b376b35734f0513a7fc66c574c8fce225852f47"} Nov 24 14:26:10 crc kubenswrapper[4824]: I1124 14:26:10.788732 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:26:10 crc kubenswrapper[4824]: I1124 14:26:10.789490 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:26:11 crc kubenswrapper[4824]: I1124 14:26:11.449607 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sh2g" event={"ID":"ff4d9122-d458-4f03-bb07-9c0a39833b2c","Type":"ContainerStarted","Data":"6a9882b5135d548f5743712d90327b78eeffdbbc43bded217abab03bc012b33b"} Nov 24 14:26:14 crc kubenswrapper[4824]: I1124 14:26:14.477100 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerID="6a9882b5135d548f5743712d90327b78eeffdbbc43bded217abab03bc012b33b" exitCode=0 Nov 24 14:26:14 crc kubenswrapper[4824]: I1124 14:26:14.477189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sh2g" event={"ID":"ff4d9122-d458-4f03-bb07-9c0a39833b2c","Type":"ContainerDied","Data":"6a9882b5135d548f5743712d90327b78eeffdbbc43bded217abab03bc012b33b"} Nov 24 14:26:15 crc kubenswrapper[4824]: I1124 14:26:15.489145 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sh2g" event={"ID":"ff4d9122-d458-4f03-bb07-9c0a39833b2c","Type":"ContainerStarted","Data":"b4203066ce83b7f09372d3ff738b84798a3b7a83a0039bafab5e8c267c00ee8a"} Nov 24 14:26:15 crc kubenswrapper[4824]: I1124 14:26:15.521369 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9sh2g" podStartSLOduration=2.944614734 podStartE2EDuration="9.521352573s" podCreationTimestamp="2025-11-24 14:26:06 +0000 UTC" firstStartedPulling="2025-11-24 14:26:08.403284816 +0000 UTC m=+4310.042824126" lastFinishedPulling="2025-11-24 14:26:14.980022655 +0000 UTC m=+4316.619561965" observedRunningTime="2025-11-24 14:26:15.520471671 +0000 UTC m=+4317.160010981" watchObservedRunningTime="2025-11-24 14:26:15.521352573 +0000 UTC m=+4317.160891883" Nov 24 14:26:16 crc kubenswrapper[4824]: I1124 14:26:16.653938 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:16 crc kubenswrapper[4824]: I1124 14:26:16.655555 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:17 crc kubenswrapper[4824]: I1124 14:26:17.710785 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9sh2g" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerName="registry-server" probeResult="failure" output=< Nov 24 14:26:17 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:26:17 crc kubenswrapper[4824]: > Nov 24 14:26:25 crc kubenswrapper[4824]: I1124 14:26:25.345163 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-8h66q_91b65029-b851-4317-a1fc-26b30a5c8734/control-plane-machine-set-operator/0.log" Nov 24 14:26:25 crc kubenswrapper[4824]: I1124 14:26:25.483474 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5r76_8a59be8e-4194-4e1c-89bd-3136366ed13b/kube-rbac-proxy/0.log" Nov 24 14:26:25 crc kubenswrapper[4824]: I1124 14:26:25.576790 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5r76_8a59be8e-4194-4e1c-89bd-3136366ed13b/machine-api-operator/0.log" Nov 24 14:26:26 crc kubenswrapper[4824]: I1124 14:26:26.697455 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:26 crc kubenswrapper[4824]: I1124 14:26:26.743011 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:26 crc kubenswrapper[4824]: I1124 14:26:26.935436 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9sh2g"] Nov 24 14:26:28 crc kubenswrapper[4824]: I1124 14:26:28.602331 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9sh2g" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerName="registry-server" containerID="cri-o://b4203066ce83b7f09372d3ff738b84798a3b7a83a0039bafab5e8c267c00ee8a" gracePeriod=2 Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.616258 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerID="b4203066ce83b7f09372d3ff738b84798a3b7a83a0039bafab5e8c267c00ee8a" exitCode=0 Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.616464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sh2g" event={"ID":"ff4d9122-d458-4f03-bb07-9c0a39833b2c","Type":"ContainerDied","Data":"b4203066ce83b7f09372d3ff738b84798a3b7a83a0039bafab5e8c267c00ee8a"} Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.730186 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.869598 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-catalog-content\") pod \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.869767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-utilities\") pod \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.869837 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85j92\" (UniqueName: \"kubernetes.io/projected/ff4d9122-d458-4f03-bb07-9c0a39833b2c-kube-api-access-85j92\") pod \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\" (UID: \"ff4d9122-d458-4f03-bb07-9c0a39833b2c\") " Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.870561 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-utilities" (OuterVolumeSpecName: "utilities") pod "ff4d9122-d458-4f03-bb07-9c0a39833b2c" (UID: "ff4d9122-d458-4f03-bb07-9c0a39833b2c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.876957 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff4d9122-d458-4f03-bb07-9c0a39833b2c-kube-api-access-85j92" (OuterVolumeSpecName: "kube-api-access-85j92") pod "ff4d9122-d458-4f03-bb07-9c0a39833b2c" (UID: "ff4d9122-d458-4f03-bb07-9c0a39833b2c"). InnerVolumeSpecName "kube-api-access-85j92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.937561 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff4d9122-d458-4f03-bb07-9c0a39833b2c" (UID: "ff4d9122-d458-4f03-bb07-9c0a39833b2c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.971984 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.972050 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85j92\" (UniqueName: \"kubernetes.io/projected/ff4d9122-d458-4f03-bb07-9c0a39833b2c-kube-api-access-85j92\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:29 crc kubenswrapper[4824]: I1124 14:26:29.972070 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4d9122-d458-4f03-bb07-9c0a39833b2c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:30 crc kubenswrapper[4824]: I1124 14:26:30.630618 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sh2g" event={"ID":"ff4d9122-d458-4f03-bb07-9c0a39833b2c","Type":"ContainerDied","Data":"57d44192242257142864aa5bdaffc7c2e49bbdb8b05feff2bd7f65a51059b76c"} Nov 24 14:26:30 crc kubenswrapper[4824]: I1124 14:26:30.630673 4824 scope.go:117] "RemoveContainer" containerID="b4203066ce83b7f09372d3ff738b84798a3b7a83a0039bafab5e8c267c00ee8a" Nov 24 14:26:30 crc kubenswrapper[4824]: I1124 14:26:30.632937 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9sh2g" Nov 24 14:26:30 crc kubenswrapper[4824]: I1124 14:26:30.659470 4824 scope.go:117] "RemoveContainer" containerID="6a9882b5135d548f5743712d90327b78eeffdbbc43bded217abab03bc012b33b" Nov 24 14:26:30 crc kubenswrapper[4824]: I1124 14:26:30.688731 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9sh2g"] Nov 24 14:26:30 crc kubenswrapper[4824]: I1124 14:26:30.702074 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9sh2g"] Nov 24 14:26:30 crc kubenswrapper[4824]: I1124 14:26:30.794876 4824 scope.go:117] "RemoveContainer" containerID="a9808b1bc64d2af944468d2b1b376b35734f0513a7fc66c574c8fce225852f47" Nov 24 14:26:31 crc kubenswrapper[4824]: I1124 14:26:31.020898 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" path="/var/lib/kubelet/pods/ff4d9122-d458-4f03-bb07-9c0a39833b2c/volumes" Nov 24 14:26:39 crc kubenswrapper[4824]: I1124 14:26:39.486618 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-8mhb8_121b83e4-cc24-4e97-831b-c05b2dce0aca/cert-manager-controller/0.log" Nov 24 14:26:39 crc kubenswrapper[4824]: I1124 14:26:39.496499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-4hmfd_db3d2465-e3f8-434a-a279-9de75d0f2bbc/cert-manager-cainjector/0.log" Nov 24 14:26:39 crc kubenswrapper[4824]: I1124 14:26:39.676330 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-p8wjb_c65d2285-3c4b-407b-937c-53a7d1302a8a/cert-manager-webhook/0.log" Nov 24 14:26:40 crc kubenswrapper[4824]: I1124 14:26:40.788379 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:26:40 crc kubenswrapper[4824]: I1124 14:26:40.788441 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:26:52 crc kubenswrapper[4824]: I1124 14:26:52.205217 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-7sv5s_a252dabf-79e7-442e-990c-3f15bc47d536/nmstate-console-plugin/0.log" Nov 24 14:26:52 crc kubenswrapper[4824]: I1124 14:26:52.418934 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gw4fj_1e8318a0-5501-47b7-836a-4d650de94219/nmstate-handler/0.log" Nov 24 14:26:52 crc kubenswrapper[4824]: I1124 14:26:52.501003 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-gb5z5_2717ebec-b84b-4d07-a1ac-67209e7e5e32/kube-rbac-proxy/0.log" Nov 24 14:26:52 crc kubenswrapper[4824]: I1124 14:26:52.592168 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-gb5z5_2717ebec-b84b-4d07-a1ac-67209e7e5e32/nmstate-metrics/0.log" Nov 24 14:26:52 crc kubenswrapper[4824]: I1124 14:26:52.695412 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-ghp4s_04d4e05b-9e17-473f-8d25-933b9a4ed60a/nmstate-operator/0.log" Nov 24 14:26:52 crc kubenswrapper[4824]: I1124 14:26:52.799834 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-qs6l4_8876f622-9722-4257-8526-9f1f0e6a3daa/nmstate-webhook/0.log" Nov 24 14:27:09 crc kubenswrapper[4824]: I1124 14:27:09.829225 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-m4d79_077fe5ab-ea92-4aa8-8b76-dc791ae6d41b/kube-rbac-proxy/0.log" Nov 24 14:27:09 crc kubenswrapper[4824]: I1124 14:27:09.939612 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-m4d79_077fe5ab-ea92-4aa8-8b76-dc791ae6d41b/controller/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.035648 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-frr-files/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.232134 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-frr-files/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.236875 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-reloader/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.274298 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-metrics/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.320210 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-reloader/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.501417 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-reloader/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.503446 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-frr-files/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.543261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-metrics/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.596769 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-metrics/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.764319 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-frr-files/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.787756 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.787852 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.787910 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.788710 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"186331f57cfd3cccc17004a62f407141eee208514b6cb45c3ca9fa815ac14b2c"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.788781 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://186331f57cfd3cccc17004a62f407141eee208514b6cb45c3ca9fa815ac14b2c" gracePeriod=600 Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.842136 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-metrics/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.874342 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/controller/0.log" Nov 24 14:27:10 crc kubenswrapper[4824]: I1124 14:27:10.874772 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-reloader/0.log" Nov 24 14:27:11 crc kubenswrapper[4824]: I1124 14:27:11.051488 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="186331f57cfd3cccc17004a62f407141eee208514b6cb45c3ca9fa815ac14b2c" exitCode=0 Nov 24 14:27:11 crc kubenswrapper[4824]: I1124 14:27:11.052037 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"186331f57cfd3cccc17004a62f407141eee208514b6cb45c3ca9fa815ac14b2c"} Nov 24 14:27:11 crc kubenswrapper[4824]: I1124 14:27:11.052175 4824 scope.go:117] "RemoveContainer" containerID="e714dcd2251c1b09f7680a0f986a9dbc53b6f171c0413a75e8565ab481c700c3" Nov 24 14:27:11 crc kubenswrapper[4824]: I1124 14:27:11.055031 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/frr-metrics/0.log" Nov 24 14:27:11 crc kubenswrapper[4824]: I1124 14:27:11.095967 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/kube-rbac-proxy/0.log" Nov 24 14:27:11 crc kubenswrapper[4824]: I1124 14:27:11.192190 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/kube-rbac-proxy-frr/0.log" Nov 24 14:27:11 crc kubenswrapper[4824]: I1124 14:27:11.368450 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/reloader/0.log" Nov 24 14:27:11 crc kubenswrapper[4824]: I1124 14:27:11.649183 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-2gp4q_db503b31-f22b-4b5c-a20f-970c15435851/frr-k8s-webhook-server/0.log" Nov 24 14:27:11 crc kubenswrapper[4824]: I1124 14:27:11.773317 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-b6cff747d-x5ng4_99485513-6292-40ee-aede-2ef1d8566124/manager/0.log" Nov 24 14:27:12 crc kubenswrapper[4824]: I1124 14:27:12.031795 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-ffbf5bdd6-fr567_07d6946c-c191-4a18-9903-194bdc00b4a0/webhook-server/0.log" Nov 24 14:27:12 crc kubenswrapper[4824]: I1124 14:27:12.061637 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668"} Nov 24 14:27:12 crc kubenswrapper[4824]: I1124 14:27:12.187413 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r5w24_e2bbe7bd-470b-4363-a85c-7a2084da3e4e/kube-rbac-proxy/0.log" Nov 24 14:27:12 crc kubenswrapper[4824]: I1124 14:27:12.303281 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/frr/0.log" Nov 24 14:27:12 crc kubenswrapper[4824]: I1124 14:27:12.744069 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r5w24_e2bbe7bd-470b-4363-a85c-7a2084da3e4e/speaker/0.log" Nov 24 14:27:25 crc kubenswrapper[4824]: I1124 14:27:25.562139 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/util/0.log" Nov 24 14:27:25 crc kubenswrapper[4824]: I1124 14:27:25.776834 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/util/0.log" Nov 24 14:27:25 crc kubenswrapper[4824]: I1124 14:27:25.800066 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/pull/0.log" Nov 24 14:27:25 crc kubenswrapper[4824]: I1124 14:27:25.826090 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/pull/0.log" Nov 24 14:27:26 crc kubenswrapper[4824]: I1124 14:27:26.050138 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/extract/0.log" Nov 24 14:27:26 crc kubenswrapper[4824]: I1124 14:27:26.063876 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/pull/0.log" Nov 24 14:27:26 crc kubenswrapper[4824]: I1124 14:27:26.081060 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/util/0.log" Nov 24 14:27:26 crc kubenswrapper[4824]: I1124 14:27:26.288138 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-utilities/0.log" Nov 24 14:27:26 crc kubenswrapper[4824]: I1124 14:27:26.405879 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-utilities/0.log" Nov 24 14:27:26 crc kubenswrapper[4824]: I1124 14:27:26.421494 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-content/0.log" Nov 24 14:27:26 crc kubenswrapper[4824]: I1124 14:27:26.426490 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-content/0.log" Nov 24 14:27:27 crc kubenswrapper[4824]: I1124 14:27:27.047566 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-content/0.log" Nov 24 14:27:27 crc kubenswrapper[4824]: I1124 14:27:27.087688 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-utilities/0.log" Nov 24 14:27:27 crc kubenswrapper[4824]: I1124 14:27:27.356428 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-utilities/0.log" Nov 24 14:27:27 crc kubenswrapper[4824]: I1124 14:27:27.552943 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-utilities/0.log" Nov 24 14:27:27 crc kubenswrapper[4824]: I1124 14:27:27.694365 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-content/0.log" Nov 24 14:27:27 crc kubenswrapper[4824]: I1124 14:27:27.721184 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-content/0.log" Nov 24 14:27:27 crc kubenswrapper[4824]: I1124 14:27:27.727952 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/registry-server/0.log" Nov 24 14:27:27 crc kubenswrapper[4824]: I1124 14:27:27.840854 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-utilities/0.log" Nov 24 14:27:27 crc kubenswrapper[4824]: I1124 14:27:27.876626 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-content/0.log" Nov 24 14:27:28 crc kubenswrapper[4824]: I1124 14:27:28.142750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/util/0.log" Nov 24 14:27:28 crc kubenswrapper[4824]: I1124 14:27:28.157075 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/registry-server/0.log" Nov 24 14:27:28 crc kubenswrapper[4824]: I1124 14:27:28.586043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/util/0.log" Nov 24 14:27:28 crc kubenswrapper[4824]: I1124 14:27:28.625237 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/pull/0.log" Nov 24 14:27:28 crc kubenswrapper[4824]: I1124 14:27:28.630548 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/pull/0.log" Nov 24 14:27:28 crc kubenswrapper[4824]: I1124 14:27:28.803145 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/util/0.log" Nov 24 14:27:28 crc kubenswrapper[4824]: I1124 14:27:28.829313 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/extract/0.log" Nov 24 14:27:28 crc kubenswrapper[4824]: I1124 14:27:28.842403 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/pull/0.log" Nov 24 14:27:29 crc kubenswrapper[4824]: I1124 14:27:29.004045 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qkj9f_7b3b4423-493d-4cc4-9533-d87011248427/marketplace-operator/0.log" Nov 24 14:27:29 crc kubenswrapper[4824]: I1124 14:27:29.108941 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-utilities/0.log" Nov 24 14:27:29 crc kubenswrapper[4824]: I1124 14:27:29.348351 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-content/0.log" Nov 24 14:27:29 crc kubenswrapper[4824]: I1124 14:27:29.366665 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-content/0.log" Nov 24 14:27:29 crc kubenswrapper[4824]: I1124 14:27:29.392260 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-utilities/0.log" Nov 24 14:27:29 crc kubenswrapper[4824]: I1124 14:27:29.640008 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-content/0.log" Nov 24 14:27:29 crc kubenswrapper[4824]: I1124 14:27:29.667177 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-utilities/0.log" Nov 24 14:27:29 crc kubenswrapper[4824]: I1124 14:27:29.672687 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/registry-server/0.log" Nov 24 14:27:29 crc kubenswrapper[4824]: I1124 14:27:29.821828 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6t7g9_373d11f5-b249-4054-9885-e3a9a684315d/extract-utilities/0.log" Nov 24 14:27:30 crc kubenswrapper[4824]: I1124 14:27:30.016193 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6t7g9_373d11f5-b249-4054-9885-e3a9a684315d/extract-utilities/0.log" Nov 24 14:27:30 crc kubenswrapper[4824]: I1124 14:27:30.066905 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6t7g9_373d11f5-b249-4054-9885-e3a9a684315d/extract-content/0.log" Nov 24 14:27:30 crc kubenswrapper[4824]: I1124 14:27:30.071298 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6t7g9_373d11f5-b249-4054-9885-e3a9a684315d/extract-content/0.log" Nov 24 14:27:30 crc kubenswrapper[4824]: I1124 14:27:30.239387 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6t7g9_373d11f5-b249-4054-9885-e3a9a684315d/extract-content/0.log" Nov 24 14:27:30 crc kubenswrapper[4824]: I1124 14:27:30.285460 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6t7g9_373d11f5-b249-4054-9885-e3a9a684315d/extract-utilities/0.log" Nov 24 14:27:30 crc kubenswrapper[4824]: I1124 14:27:30.608363 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6t7g9_373d11f5-b249-4054-9885-e3a9a684315d/registry-server/0.log" Nov 24 14:28:06 crc kubenswrapper[4824]: E1124 14:28:06.770976 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.188:47114->38.129.56.188:33927: write tcp 38.129.56.188:47114->38.129.56.188:33927: write: connection reset by peer Nov 24 14:29:40 crc kubenswrapper[4824]: I1124 14:29:40.787729 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:29:40 crc kubenswrapper[4824]: I1124 14:29:40.789388 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:29:55 crc kubenswrapper[4824]: I1124 14:29:55.588312 4824 generic.go:334] "Generic (PLEG): container finished" podID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" containerID="fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1" exitCode=0 Nov 24 14:29:55 crc kubenswrapper[4824]: I1124 14:29:55.588431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hk622/must-gather-dkjtr" event={"ID":"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0","Type":"ContainerDied","Data":"fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1"} Nov 24 14:29:55 crc kubenswrapper[4824]: I1124 14:29:55.589551 4824 scope.go:117] "RemoveContainer" containerID="fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1" Nov 24 14:29:55 crc kubenswrapper[4824]: I1124 14:29:55.852831 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hk622_must-gather-dkjtr_3cbb2e20-3f4d-4374-9062-c05e9b1de7c0/gather/0.log" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.155049 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw"] Nov 24 14:30:00 crc kubenswrapper[4824]: E1124 14:30:00.156153 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerName="extract-content" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.156173 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerName="extract-content" Nov 24 14:30:00 crc kubenswrapper[4824]: E1124 14:30:00.156194 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerName="registry-server" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.156204 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerName="registry-server" Nov 24 14:30:00 crc kubenswrapper[4824]: E1124 14:30:00.156215 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerName="extract-utilities" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.156223 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerName="extract-utilities" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.156466 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff4d9122-d458-4f03-bb07-9c0a39833b2c" containerName="registry-server" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.157237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.159254 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.164355 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw"] Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.174180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.205054 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnn82\" (UniqueName: \"kubernetes.io/projected/d4ed8ef8-9ad1-4385-9731-9687161d40df-kube-api-access-pnn82\") pod \"collect-profiles-29399910-wjgzw\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.205174 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4ed8ef8-9ad1-4385-9731-9687161d40df-secret-volume\") pod \"collect-profiles-29399910-wjgzw\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.205223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4ed8ef8-9ad1-4385-9731-9687161d40df-config-volume\") pod \"collect-profiles-29399910-wjgzw\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.307048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4ed8ef8-9ad1-4385-9731-9687161d40df-config-volume\") pod \"collect-profiles-29399910-wjgzw\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.307174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnn82\" (UniqueName: \"kubernetes.io/projected/d4ed8ef8-9ad1-4385-9731-9687161d40df-kube-api-access-pnn82\") pod \"collect-profiles-29399910-wjgzw\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.307229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4ed8ef8-9ad1-4385-9731-9687161d40df-secret-volume\") pod \"collect-profiles-29399910-wjgzw\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.307946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4ed8ef8-9ad1-4385-9731-9687161d40df-config-volume\") pod \"collect-profiles-29399910-wjgzw\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.320772 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4ed8ef8-9ad1-4385-9731-9687161d40df-secret-volume\") pod \"collect-profiles-29399910-wjgzw\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.323216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnn82\" (UniqueName: \"kubernetes.io/projected/d4ed8ef8-9ad1-4385-9731-9687161d40df-kube-api-access-pnn82\") pod \"collect-profiles-29399910-wjgzw\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: I1124 14:30:00.487905 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:00 crc kubenswrapper[4824]: W1124 14:30:00.997082 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4ed8ef8_9ad1_4385_9731_9687161d40df.slice/crio-50fac9ed2ba576c47985cbd5eac3d1513db9917845331593c03e0ace44247570 WatchSource:0}: Error finding container 50fac9ed2ba576c47985cbd5eac3d1513db9917845331593c03e0ace44247570: Status 404 returned error can't find the container with id 50fac9ed2ba576c47985cbd5eac3d1513db9917845331593c03e0ace44247570 Nov 24 14:30:01 crc kubenswrapper[4824]: I1124 14:30:01.026662 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw"] Nov 24 14:30:01 crc kubenswrapper[4824]: I1124 14:30:01.653108 4824 generic.go:334] "Generic (PLEG): container finished" podID="d4ed8ef8-9ad1-4385-9731-9687161d40df" containerID="81c39d3466e0d14ffd73f1be218a2ba6cca3eabf47d6a929b3bcd178bf290979" exitCode=0 Nov 24 14:30:01 crc kubenswrapper[4824]: I1124 14:30:01.653584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" event={"ID":"d4ed8ef8-9ad1-4385-9731-9687161d40df","Type":"ContainerDied","Data":"81c39d3466e0d14ffd73f1be218a2ba6cca3eabf47d6a929b3bcd178bf290979"} Nov 24 14:30:01 crc kubenswrapper[4824]: I1124 14:30:01.654403 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" event={"ID":"d4ed8ef8-9ad1-4385-9731-9687161d40df","Type":"ContainerStarted","Data":"50fac9ed2ba576c47985cbd5eac3d1513db9917845331593c03e0ace44247570"} Nov 24 14:30:01 crc kubenswrapper[4824]: I1124 14:30:01.970086 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xm5c6"] Nov 24 14:30:01 crc kubenswrapper[4824]: I1124 14:30:01.977337 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.001651 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xm5c6"] Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.046776 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxrcg\" (UniqueName: \"kubernetes.io/projected/9f0b9253-e525-4c45-9f90-036eceb195d1-kube-api-access-gxrcg\") pod \"redhat-operators-xm5c6\" (UID: \"9f0b9253-e525-4c45-9f90-036eceb195d1\") " pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.046871 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f0b9253-e525-4c45-9f90-036eceb195d1-catalog-content\") pod \"redhat-operators-xm5c6\" (UID: \"9f0b9253-e525-4c45-9f90-036eceb195d1\") " pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.047571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f0b9253-e525-4c45-9f90-036eceb195d1-utilities\") pod \"redhat-operators-xm5c6\" (UID: \"9f0b9253-e525-4c45-9f90-036eceb195d1\") " pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.149138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f0b9253-e525-4c45-9f90-036eceb195d1-utilities\") pod \"redhat-operators-xm5c6\" (UID: \"9f0b9253-e525-4c45-9f90-036eceb195d1\") " pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.149212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxrcg\" (UniqueName: \"kubernetes.io/projected/9f0b9253-e525-4c45-9f90-036eceb195d1-kube-api-access-gxrcg\") pod \"redhat-operators-xm5c6\" (UID: \"9f0b9253-e525-4c45-9f90-036eceb195d1\") " pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.149251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f0b9253-e525-4c45-9f90-036eceb195d1-catalog-content\") pod \"redhat-operators-xm5c6\" (UID: \"9f0b9253-e525-4c45-9f90-036eceb195d1\") " pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.149642 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f0b9253-e525-4c45-9f90-036eceb195d1-utilities\") pod \"redhat-operators-xm5c6\" (UID: \"9f0b9253-e525-4c45-9f90-036eceb195d1\") " pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.149726 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f0b9253-e525-4c45-9f90-036eceb195d1-catalog-content\") pod \"redhat-operators-xm5c6\" (UID: \"9f0b9253-e525-4c45-9f90-036eceb195d1\") " pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.182469 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxrcg\" (UniqueName: \"kubernetes.io/projected/9f0b9253-e525-4c45-9f90-036eceb195d1-kube-api-access-gxrcg\") pod \"redhat-operators-xm5c6\" (UID: \"9f0b9253-e525-4c45-9f90-036eceb195d1\") " pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.308633 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:02 crc kubenswrapper[4824]: I1124 14:30:02.858322 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xm5c6"] Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.090834 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.166261 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnn82\" (UniqueName: \"kubernetes.io/projected/d4ed8ef8-9ad1-4385-9731-9687161d40df-kube-api-access-pnn82\") pod \"d4ed8ef8-9ad1-4385-9731-9687161d40df\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.166318 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4ed8ef8-9ad1-4385-9731-9687161d40df-config-volume\") pod \"d4ed8ef8-9ad1-4385-9731-9687161d40df\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.166430 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4ed8ef8-9ad1-4385-9731-9687161d40df-secret-volume\") pod \"d4ed8ef8-9ad1-4385-9731-9687161d40df\" (UID: \"d4ed8ef8-9ad1-4385-9731-9687161d40df\") " Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.169530 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4ed8ef8-9ad1-4385-9731-9687161d40df-config-volume" (OuterVolumeSpecName: "config-volume") pod "d4ed8ef8-9ad1-4385-9731-9687161d40df" (UID: "d4ed8ef8-9ad1-4385-9731-9687161d40df"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.175530 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4ed8ef8-9ad1-4385-9731-9687161d40df-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d4ed8ef8-9ad1-4385-9731-9687161d40df" (UID: "d4ed8ef8-9ad1-4385-9731-9687161d40df"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.178104 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4ed8ef8-9ad1-4385-9731-9687161d40df-kube-api-access-pnn82" (OuterVolumeSpecName: "kube-api-access-pnn82") pod "d4ed8ef8-9ad1-4385-9731-9687161d40df" (UID: "d4ed8ef8-9ad1-4385-9731-9687161d40df"). InnerVolumeSpecName "kube-api-access-pnn82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.268041 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnn82\" (UniqueName: \"kubernetes.io/projected/d4ed8ef8-9ad1-4385-9731-9687161d40df-kube-api-access-pnn82\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.268070 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4ed8ef8-9ad1-4385-9731-9687161d40df-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.268079 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4ed8ef8-9ad1-4385-9731-9687161d40df-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.671405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" event={"ID":"d4ed8ef8-9ad1-4385-9731-9687161d40df","Type":"ContainerDied","Data":"50fac9ed2ba576c47985cbd5eac3d1513db9917845331593c03e0ace44247570"} Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.671766 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50fac9ed2ba576c47985cbd5eac3d1513db9917845331593c03e0ace44247570" Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.671450 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-wjgzw" Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.673894 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f0b9253-e525-4c45-9f90-036eceb195d1" containerID="8d8365f8ed9eb7f2a6dc46fb36067ca3c7665e57aa6905cb10f12770b0c42aa9" exitCode=0 Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.673937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xm5c6" event={"ID":"9f0b9253-e525-4c45-9f90-036eceb195d1","Type":"ContainerDied","Data":"8d8365f8ed9eb7f2a6dc46fb36067ca3c7665e57aa6905cb10f12770b0c42aa9"} Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.673967 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xm5c6" event={"ID":"9f0b9253-e525-4c45-9f90-036eceb195d1","Type":"ContainerStarted","Data":"dc9e90ccb351acc515e5ea37085214d35131d074c3259ecf4bcfa25dcb6944e4"} Nov 24 14:30:03 crc kubenswrapper[4824]: I1124 14:30:03.676623 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:30:04 crc kubenswrapper[4824]: I1124 14:30:04.172104 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj"] Nov 24 14:30:04 crc kubenswrapper[4824]: I1124 14:30:04.179630 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-x95hj"] Nov 24 14:30:04 crc kubenswrapper[4824]: I1124 14:30:04.728566 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hk622/must-gather-dkjtr"] Nov 24 14:30:04 crc kubenswrapper[4824]: I1124 14:30:04.729199 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-hk622/must-gather-dkjtr" podUID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" containerName="copy" containerID="cri-o://036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960" gracePeriod=2 Nov 24 14:30:04 crc kubenswrapper[4824]: I1124 14:30:04.738599 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hk622/must-gather-dkjtr"] Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.073170 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f59d69e2-3080-431f-8b9b-c775a6397109" path="/var/lib/kubelet/pods/f59d69e2-3080-431f-8b9b-c775a6397109/volumes" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.216288 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hk622_must-gather-dkjtr_3cbb2e20-3f4d-4374-9062-c05e9b1de7c0/copy/0.log" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.216902 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.311221 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jjn8\" (UniqueName: \"kubernetes.io/projected/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-kube-api-access-8jjn8\") pod \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\" (UID: \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\") " Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.311281 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-must-gather-output\") pod \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\" (UID: \"3cbb2e20-3f4d-4374-9062-c05e9b1de7c0\") " Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.318849 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-kube-api-access-8jjn8" (OuterVolumeSpecName: "kube-api-access-8jjn8") pod "3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" (UID: "3cbb2e20-3f4d-4374-9062-c05e9b1de7c0"). InnerVolumeSpecName "kube-api-access-8jjn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.413887 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jjn8\" (UniqueName: \"kubernetes.io/projected/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-kube-api-access-8jjn8\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.482504 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" (UID: "3cbb2e20-3f4d-4374-9062-c05e9b1de7c0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.516129 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.692646 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hk622_must-gather-dkjtr_3cbb2e20-3f4d-4374-9062-c05e9b1de7c0/copy/0.log" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.693188 4824 generic.go:334] "Generic (PLEG): container finished" podID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" containerID="036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960" exitCode=143 Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.693240 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hk622/must-gather-dkjtr" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.693267 4824 scope.go:117] "RemoveContainer" containerID="036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960" Nov 24 14:30:05 crc kubenswrapper[4824]: I1124 14:30:05.715918 4824 scope.go:117] "RemoveContainer" containerID="fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1" Nov 24 14:30:06 crc kubenswrapper[4824]: I1124 14:30:06.088645 4824 scope.go:117] "RemoveContainer" containerID="036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960" Nov 24 14:30:06 crc kubenswrapper[4824]: E1124 14:30:06.094278 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960\": container with ID starting with 036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960 not found: ID does not exist" containerID="036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960" Nov 24 14:30:06 crc kubenswrapper[4824]: I1124 14:30:06.094321 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960"} err="failed to get container status \"036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960\": rpc error: code = NotFound desc = could not find container \"036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960\": container with ID starting with 036f35b2b3eeee0bd4e2d74dec36fbfbb5ec88527ab369096994b7df4fee7960 not found: ID does not exist" Nov 24 14:30:06 crc kubenswrapper[4824]: I1124 14:30:06.094349 4824 scope.go:117] "RemoveContainer" containerID="fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1" Nov 24 14:30:06 crc kubenswrapper[4824]: E1124 14:30:06.094679 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1\": container with ID starting with fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1 not found: ID does not exist" containerID="fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1" Nov 24 14:30:06 crc kubenswrapper[4824]: I1124 14:30:06.094742 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1"} err="failed to get container status \"fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1\": rpc error: code = NotFound desc = could not find container \"fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1\": container with ID starting with fd0e1ee756d42329790afb466d0f003336aee97833b32700fd5f2d720b7f65c1 not found: ID does not exist" Nov 24 14:30:07 crc kubenswrapper[4824]: I1124 14:30:07.022957 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" path="/var/lib/kubelet/pods/3cbb2e20-3f4d-4374-9062-c05e9b1de7c0/volumes" Nov 24 14:30:10 crc kubenswrapper[4824]: I1124 14:30:10.793350 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:30:10 crc kubenswrapper[4824]: I1124 14:30:10.793841 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:30:14 crc kubenswrapper[4824]: I1124 14:30:14.797153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xm5c6" event={"ID":"9f0b9253-e525-4c45-9f90-036eceb195d1","Type":"ContainerStarted","Data":"469bd31442468e8d3eb8305e33b6168164d55c6a712727ae091603bb40227f1d"} Nov 24 14:30:16 crc kubenswrapper[4824]: I1124 14:30:16.816161 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f0b9253-e525-4c45-9f90-036eceb195d1" containerID="469bd31442468e8d3eb8305e33b6168164d55c6a712727ae091603bb40227f1d" exitCode=0 Nov 24 14:30:16 crc kubenswrapper[4824]: I1124 14:30:16.816354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xm5c6" event={"ID":"9f0b9253-e525-4c45-9f90-036eceb195d1","Type":"ContainerDied","Data":"469bd31442468e8d3eb8305e33b6168164d55c6a712727ae091603bb40227f1d"} Nov 24 14:30:20 crc kubenswrapper[4824]: I1124 14:30:20.148933 4824 trace.go:236] Trace[1011377417]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/redhat-operators-xm5c6" (24-Nov-2025 14:30:19.062) (total time: 1085ms): Nov 24 14:30:20 crc kubenswrapper[4824]: Trace[1011377417]: [1.085740739s] [1.085740739s] END Nov 24 14:30:20 crc kubenswrapper[4824]: I1124 14:30:20.858719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xm5c6" event={"ID":"9f0b9253-e525-4c45-9f90-036eceb195d1","Type":"ContainerStarted","Data":"410b88ab70d67c33136fb88a82d130fa6a9ba1a561ba71961b50f8ca14694e3a"} Nov 24 14:30:20 crc kubenswrapper[4824]: I1124 14:30:20.885897 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xm5c6" podStartSLOduration=3.30977418 podStartE2EDuration="19.885877532s" podCreationTimestamp="2025-11-24 14:30:01 +0000 UTC" firstStartedPulling="2025-11-24 14:30:03.676366142 +0000 UTC m=+4545.315905452" lastFinishedPulling="2025-11-24 14:30:20.252469454 +0000 UTC m=+4561.892008804" observedRunningTime="2025-11-24 14:30:20.885738639 +0000 UTC m=+4562.525277959" watchObservedRunningTime="2025-11-24 14:30:20.885877532 +0000 UTC m=+4562.525416852" Nov 24 14:30:22 crc kubenswrapper[4824]: I1124 14:30:22.309896 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:22 crc kubenswrapper[4824]: I1124 14:30:22.311113 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:23 crc kubenswrapper[4824]: I1124 14:30:23.361389 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xm5c6" podUID="9f0b9253-e525-4c45-9f90-036eceb195d1" containerName="registry-server" probeResult="failure" output=< Nov 24 14:30:23 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:30:23 crc kubenswrapper[4824]: > Nov 24 14:30:25 crc kubenswrapper[4824]: I1124 14:30:25.623647 4824 scope.go:117] "RemoveContainer" containerID="d1f7b025fb9f3898fbe0d0c327b1877f2a52caca91c9e7f7990ae00f9b428d55" Nov 24 14:30:33 crc kubenswrapper[4824]: I1124 14:30:33.402408 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xm5c6" podUID="9f0b9253-e525-4c45-9f90-036eceb195d1" containerName="registry-server" probeResult="failure" output=< Nov 24 14:30:33 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:30:33 crc kubenswrapper[4824]: > Nov 24 14:30:40 crc kubenswrapper[4824]: I1124 14:30:40.787683 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:30:40 crc kubenswrapper[4824]: I1124 14:30:40.788324 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:30:40 crc kubenswrapper[4824]: I1124 14:30:40.788369 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 14:30:40 crc kubenswrapper[4824]: I1124 14:30:40.789155 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:30:40 crc kubenswrapper[4824]: I1124 14:30:40.789209 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" gracePeriod=600 Nov 24 14:30:40 crc kubenswrapper[4824]: E1124 14:30:40.918011 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:30:41 crc kubenswrapper[4824]: I1124 14:30:41.018923 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" exitCode=0 Nov 24 14:30:41 crc kubenswrapper[4824]: I1124 14:30:41.023431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668"} Nov 24 14:30:41 crc kubenswrapper[4824]: I1124 14:30:41.023479 4824 scope.go:117] "RemoveContainer" containerID="186331f57cfd3cccc17004a62f407141eee208514b6cb45c3ca9fa815ac14b2c" Nov 24 14:30:41 crc kubenswrapper[4824]: I1124 14:30:41.024183 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:30:41 crc kubenswrapper[4824]: E1124 14:30:41.024460 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:30:42 crc kubenswrapper[4824]: I1124 14:30:42.358613 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:42 crc kubenswrapper[4824]: I1124 14:30:42.413972 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xm5c6" Nov 24 14:30:42 crc kubenswrapper[4824]: I1124 14:30:42.501040 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xm5c6"] Nov 24 14:30:42 crc kubenswrapper[4824]: I1124 14:30:42.590234 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6t7g9"] Nov 24 14:30:42 crc kubenswrapper[4824]: I1124 14:30:42.590478 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6t7g9" podUID="373d11f5-b249-4054-9885-e3a9a684315d" containerName="registry-server" containerID="cri-o://160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a" gracePeriod=2 Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.036183 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.049640 4824 generic.go:334] "Generic (PLEG): container finished" podID="373d11f5-b249-4054-9885-e3a9a684315d" containerID="160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a" exitCode=0 Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.049723 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6t7g9" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.049789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6t7g9" event={"ID":"373d11f5-b249-4054-9885-e3a9a684315d","Type":"ContainerDied","Data":"160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a"} Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.049862 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6t7g9" event={"ID":"373d11f5-b249-4054-9885-e3a9a684315d","Type":"ContainerDied","Data":"f2ad860e8259a4b8298869360fb7be0d1dd34672f870b9a1459db5a21961f278"} Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.049883 4824 scope.go:117] "RemoveContainer" containerID="160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.095818 4824 scope.go:117] "RemoveContainer" containerID="650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.122699 4824 scope.go:117] "RemoveContainer" containerID="0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.145026 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9tpx\" (UniqueName: \"kubernetes.io/projected/373d11f5-b249-4054-9885-e3a9a684315d-kube-api-access-x9tpx\") pod \"373d11f5-b249-4054-9885-e3a9a684315d\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.145518 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-catalog-content\") pod \"373d11f5-b249-4054-9885-e3a9a684315d\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.145634 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-utilities\") pod \"373d11f5-b249-4054-9885-e3a9a684315d\" (UID: \"373d11f5-b249-4054-9885-e3a9a684315d\") " Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.149259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-utilities" (OuterVolumeSpecName: "utilities") pod "373d11f5-b249-4054-9885-e3a9a684315d" (UID: "373d11f5-b249-4054-9885-e3a9a684315d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.151658 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373d11f5-b249-4054-9885-e3a9a684315d-kube-api-access-x9tpx" (OuterVolumeSpecName: "kube-api-access-x9tpx") pod "373d11f5-b249-4054-9885-e3a9a684315d" (UID: "373d11f5-b249-4054-9885-e3a9a684315d"). InnerVolumeSpecName "kube-api-access-x9tpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.224708 4824 scope.go:117] "RemoveContainer" containerID="160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a" Nov 24 14:30:43 crc kubenswrapper[4824]: E1124 14:30:43.225269 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a\": container with ID starting with 160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a not found: ID does not exist" containerID="160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.225313 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a"} err="failed to get container status \"160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a\": rpc error: code = NotFound desc = could not find container \"160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a\": container with ID starting with 160829db2415fd29618adcc087896f4d3e471152ffb0a3c2dabc450d9ce8ea3a not found: ID does not exist" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.225344 4824 scope.go:117] "RemoveContainer" containerID="650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212" Nov 24 14:30:43 crc kubenswrapper[4824]: E1124 14:30:43.226022 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212\": container with ID starting with 650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212 not found: ID does not exist" containerID="650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.226140 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212"} err="failed to get container status \"650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212\": rpc error: code = NotFound desc = could not find container \"650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212\": container with ID starting with 650be30198c02bd02417e0fadb559a6dd911b92546e655e5b673730f9c5f5212 not found: ID does not exist" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.226242 4824 scope.go:117] "RemoveContainer" containerID="0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a" Nov 24 14:30:43 crc kubenswrapper[4824]: E1124 14:30:43.226638 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a\": container with ID starting with 0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a not found: ID does not exist" containerID="0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.226673 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a"} err="failed to get container status \"0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a\": rpc error: code = NotFound desc = could not find container \"0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a\": container with ID starting with 0ca64481b112c99bf6f9088961099183fe55c049d62a88167fb823323bd4cc4a not found: ID does not exist" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.251276 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.251317 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9tpx\" (UniqueName: \"kubernetes.io/projected/373d11f5-b249-4054-9885-e3a9a684315d-kube-api-access-x9tpx\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.279536 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "373d11f5-b249-4054-9885-e3a9a684315d" (UID: "373d11f5-b249-4054-9885-e3a9a684315d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.352845 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d11f5-b249-4054-9885-e3a9a684315d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.385947 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6t7g9"] Nov 24 14:30:43 crc kubenswrapper[4824]: I1124 14:30:43.392755 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6t7g9"] Nov 24 14:30:45 crc kubenswrapper[4824]: I1124 14:30:45.020934 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="373d11f5-b249-4054-9885-e3a9a684315d" path="/var/lib/kubelet/pods/373d11f5-b249-4054-9885-e3a9a684315d/volumes" Nov 24 14:30:54 crc kubenswrapper[4824]: I1124 14:30:54.011523 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:30:54 crc kubenswrapper[4824]: E1124 14:30:54.012713 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:31:05 crc kubenswrapper[4824]: I1124 14:31:05.010375 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:31:05 crc kubenswrapper[4824]: E1124 14:31:05.011045 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:31:17 crc kubenswrapper[4824]: I1124 14:31:17.011415 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:31:17 crc kubenswrapper[4824]: E1124 14:31:17.012604 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:31:25 crc kubenswrapper[4824]: I1124 14:31:25.720290 4824 scope.go:117] "RemoveContainer" containerID="396e5d811bfd763d037363bcb502dd5355911ad5874fc306936b32734a95fb50" Nov 24 14:31:31 crc kubenswrapper[4824]: I1124 14:31:31.011661 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:31:31 crc kubenswrapper[4824]: E1124 14:31:31.012656 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:31:42 crc kubenswrapper[4824]: I1124 14:31:42.010953 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:31:42 crc kubenswrapper[4824]: E1124 14:31:42.011915 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:31:53 crc kubenswrapper[4824]: I1124 14:31:53.010088 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:31:53 crc kubenswrapper[4824]: E1124 14:31:53.010954 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:32:08 crc kubenswrapper[4824]: I1124 14:32:08.011278 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:32:08 crc kubenswrapper[4824]: E1124 14:32:08.012309 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:32:20 crc kubenswrapper[4824]: I1124 14:32:20.009958 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:32:20 crc kubenswrapper[4824]: E1124 14:32:20.010584 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:32:33 crc kubenswrapper[4824]: I1124 14:32:33.012046 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:32:33 crc kubenswrapper[4824]: E1124 14:32:33.013574 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.472481 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9hlbw/must-gather-m4hwq"] Nov 24 14:32:37 crc kubenswrapper[4824]: E1124 14:32:37.473565 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373d11f5-b249-4054-9885-e3a9a684315d" containerName="extract-content" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.473583 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="373d11f5-b249-4054-9885-e3a9a684315d" containerName="extract-content" Nov 24 14:32:37 crc kubenswrapper[4824]: E1124 14:32:37.473603 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373d11f5-b249-4054-9885-e3a9a684315d" containerName="extract-utilities" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.473616 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="373d11f5-b249-4054-9885-e3a9a684315d" containerName="extract-utilities" Nov 24 14:32:37 crc kubenswrapper[4824]: E1124 14:32:37.473641 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" containerName="copy" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.473652 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" containerName="copy" Nov 24 14:32:37 crc kubenswrapper[4824]: E1124 14:32:37.473675 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4ed8ef8-9ad1-4385-9731-9687161d40df" containerName="collect-profiles" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.473686 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4ed8ef8-9ad1-4385-9731-9687161d40df" containerName="collect-profiles" Nov 24 14:32:37 crc kubenswrapper[4824]: E1124 14:32:37.473714 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373d11f5-b249-4054-9885-e3a9a684315d" containerName="registry-server" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.473723 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="373d11f5-b249-4054-9885-e3a9a684315d" containerName="registry-server" Nov 24 14:32:37 crc kubenswrapper[4824]: E1124 14:32:37.473751 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" containerName="gather" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.473758 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" containerName="gather" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.473995 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="373d11f5-b249-4054-9885-e3a9a684315d" containerName="registry-server" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.474006 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4ed8ef8-9ad1-4385-9731-9687161d40df" containerName="collect-profiles" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.474029 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" containerName="copy" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.474052 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cbb2e20-3f4d-4374-9062-c05e9b1de7c0" containerName="gather" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.475277 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.482652 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-9hlbw"/"default-dockercfg-dpj4k" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.482652 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9hlbw"/"openshift-service-ca.crt" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.483521 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9hlbw"/"kube-root-ca.crt" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.544179 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9hlbw/must-gather-m4hwq"] Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.612963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5689a9fe-8af0-47a5-947a-b191a6954d3f-must-gather-output\") pod \"must-gather-m4hwq\" (UID: \"5689a9fe-8af0-47a5-947a-b191a6954d3f\") " pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.613059 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hgb8\" (UniqueName: \"kubernetes.io/projected/5689a9fe-8af0-47a5-947a-b191a6954d3f-kube-api-access-6hgb8\") pod \"must-gather-m4hwq\" (UID: \"5689a9fe-8af0-47a5-947a-b191a6954d3f\") " pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.715600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5689a9fe-8af0-47a5-947a-b191a6954d3f-must-gather-output\") pod \"must-gather-m4hwq\" (UID: \"5689a9fe-8af0-47a5-947a-b191a6954d3f\") " pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.715684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hgb8\" (UniqueName: \"kubernetes.io/projected/5689a9fe-8af0-47a5-947a-b191a6954d3f-kube-api-access-6hgb8\") pod \"must-gather-m4hwq\" (UID: \"5689a9fe-8af0-47a5-947a-b191a6954d3f\") " pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:32:37 crc kubenswrapper[4824]: I1124 14:32:37.716469 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5689a9fe-8af0-47a5-947a-b191a6954d3f-must-gather-output\") pod \"must-gather-m4hwq\" (UID: \"5689a9fe-8af0-47a5-947a-b191a6954d3f\") " pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:32:38 crc kubenswrapper[4824]: I1124 14:32:38.079838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hgb8\" (UniqueName: \"kubernetes.io/projected/5689a9fe-8af0-47a5-947a-b191a6954d3f-kube-api-access-6hgb8\") pod \"must-gather-m4hwq\" (UID: \"5689a9fe-8af0-47a5-947a-b191a6954d3f\") " pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:32:38 crc kubenswrapper[4824]: I1124 14:32:38.095669 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:32:38 crc kubenswrapper[4824]: I1124 14:32:38.618482 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9hlbw/must-gather-m4hwq"] Nov 24 14:32:39 crc kubenswrapper[4824]: I1124 14:32:39.440221 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" event={"ID":"5689a9fe-8af0-47a5-947a-b191a6954d3f","Type":"ContainerStarted","Data":"5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1"} Nov 24 14:32:39 crc kubenswrapper[4824]: I1124 14:32:39.445452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" event={"ID":"5689a9fe-8af0-47a5-947a-b191a6954d3f","Type":"ContainerStarted","Data":"4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82"} Nov 24 14:32:39 crc kubenswrapper[4824]: I1124 14:32:39.445493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" event={"ID":"5689a9fe-8af0-47a5-947a-b191a6954d3f","Type":"ContainerStarted","Data":"92468ed30dff45a50ee13eb4ec83494f39bb4aa5c342c59f055510b3ae015a63"} Nov 24 14:32:39 crc kubenswrapper[4824]: I1124 14:32:39.476949 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" podStartSLOduration=2.476923964 podStartE2EDuration="2.476923964s" podCreationTimestamp="2025-11-24 14:32:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:32:39.463706105 +0000 UTC m=+4701.103245445" watchObservedRunningTime="2025-11-24 14:32:39.476923964 +0000 UTC m=+4701.116463474" Nov 24 14:32:42 crc kubenswrapper[4824]: I1124 14:32:42.872346 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9hlbw/crc-debug-wzj6c"] Nov 24 14:32:42 crc kubenswrapper[4824]: I1124 14:32:42.873994 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:32:43 crc kubenswrapper[4824]: I1124 14:32:43.013272 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e993edd4-f924-4ccc-b818-6718b9c8d7b3-host\") pod \"crc-debug-wzj6c\" (UID: \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\") " pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:32:43 crc kubenswrapper[4824]: I1124 14:32:43.013525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-568gf\" (UniqueName: \"kubernetes.io/projected/e993edd4-f924-4ccc-b818-6718b9c8d7b3-kube-api-access-568gf\") pod \"crc-debug-wzj6c\" (UID: \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\") " pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:32:43 crc kubenswrapper[4824]: I1124 14:32:43.115150 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e993edd4-f924-4ccc-b818-6718b9c8d7b3-host\") pod \"crc-debug-wzj6c\" (UID: \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\") " pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:32:43 crc kubenswrapper[4824]: I1124 14:32:43.115255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-568gf\" (UniqueName: \"kubernetes.io/projected/e993edd4-f924-4ccc-b818-6718b9c8d7b3-kube-api-access-568gf\") pod \"crc-debug-wzj6c\" (UID: \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\") " pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:32:43 crc kubenswrapper[4824]: I1124 14:32:43.115314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e993edd4-f924-4ccc-b818-6718b9c8d7b3-host\") pod \"crc-debug-wzj6c\" (UID: \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\") " pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:32:43 crc kubenswrapper[4824]: I1124 14:32:43.143837 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-568gf\" (UniqueName: \"kubernetes.io/projected/e993edd4-f924-4ccc-b818-6718b9c8d7b3-kube-api-access-568gf\") pod \"crc-debug-wzj6c\" (UID: \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\") " pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:32:43 crc kubenswrapper[4824]: I1124 14:32:43.189918 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:32:43 crc kubenswrapper[4824]: W1124 14:32:43.226787 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode993edd4_f924_4ccc_b818_6718b9c8d7b3.slice/crio-2b8f843ff2aeeaf7f92dd8ce90b04423c2624f45f2cb345fee3b2851505a7bc2 WatchSource:0}: Error finding container 2b8f843ff2aeeaf7f92dd8ce90b04423c2624f45f2cb345fee3b2851505a7bc2: Status 404 returned error can't find the container with id 2b8f843ff2aeeaf7f92dd8ce90b04423c2624f45f2cb345fee3b2851505a7bc2 Nov 24 14:32:43 crc kubenswrapper[4824]: I1124 14:32:43.475790 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" event={"ID":"e993edd4-f924-4ccc-b818-6718b9c8d7b3","Type":"ContainerStarted","Data":"2b8f843ff2aeeaf7f92dd8ce90b04423c2624f45f2cb345fee3b2851505a7bc2"} Nov 24 14:32:44 crc kubenswrapper[4824]: I1124 14:32:44.503384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" event={"ID":"e993edd4-f924-4ccc-b818-6718b9c8d7b3","Type":"ContainerStarted","Data":"6336ef52a9c45ebc6fc7203300e355a8e9fb95220b69e3f74176113cc6f4f290"} Nov 24 14:32:44 crc kubenswrapper[4824]: I1124 14:32:44.524953 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" podStartSLOduration=2.524932789 podStartE2EDuration="2.524932789s" podCreationTimestamp="2025-11-24 14:32:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:32:44.520641392 +0000 UTC m=+4706.160180722" watchObservedRunningTime="2025-11-24 14:32:44.524932789 +0000 UTC m=+4706.164472109" Nov 24 14:32:48 crc kubenswrapper[4824]: I1124 14:32:48.011060 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:32:48 crc kubenswrapper[4824]: E1124 14:32:48.012213 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:33:00 crc kubenswrapper[4824]: I1124 14:33:00.010510 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:33:00 crc kubenswrapper[4824]: E1124 14:33:00.011326 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:33:14 crc kubenswrapper[4824]: I1124 14:33:14.010444 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:33:14 crc kubenswrapper[4824]: E1124 14:33:14.012226 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:33:27 crc kubenswrapper[4824]: I1124 14:33:27.010771 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:33:27 crc kubenswrapper[4824]: E1124 14:33:27.011471 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:33:28 crc kubenswrapper[4824]: I1124 14:33:28.886608 4824 generic.go:334] "Generic (PLEG): container finished" podID="e993edd4-f924-4ccc-b818-6718b9c8d7b3" containerID="6336ef52a9c45ebc6fc7203300e355a8e9fb95220b69e3f74176113cc6f4f290" exitCode=0 Nov 24 14:33:28 crc kubenswrapper[4824]: I1124 14:33:28.886688 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" event={"ID":"e993edd4-f924-4ccc-b818-6718b9c8d7b3","Type":"ContainerDied","Data":"6336ef52a9c45ebc6fc7203300e355a8e9fb95220b69e3f74176113cc6f4f290"} Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.317774 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.354585 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9hlbw/crc-debug-wzj6c"] Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.362128 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9hlbw/crc-debug-wzj6c"] Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.431465 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-568gf\" (UniqueName: \"kubernetes.io/projected/e993edd4-f924-4ccc-b818-6718b9c8d7b3-kube-api-access-568gf\") pod \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\" (UID: \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\") " Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.431591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e993edd4-f924-4ccc-b818-6718b9c8d7b3-host\") pod \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\" (UID: \"e993edd4-f924-4ccc-b818-6718b9c8d7b3\") " Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.431705 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e993edd4-f924-4ccc-b818-6718b9c8d7b3-host" (OuterVolumeSpecName: "host") pod "e993edd4-f924-4ccc-b818-6718b9c8d7b3" (UID: "e993edd4-f924-4ccc-b818-6718b9c8d7b3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.431991 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e993edd4-f924-4ccc-b818-6718b9c8d7b3-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.437654 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e993edd4-f924-4ccc-b818-6718b9c8d7b3-kube-api-access-568gf" (OuterVolumeSpecName: "kube-api-access-568gf") pod "e993edd4-f924-4ccc-b818-6718b9c8d7b3" (UID: "e993edd4-f924-4ccc-b818-6718b9c8d7b3"). InnerVolumeSpecName "kube-api-access-568gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.534096 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-568gf\" (UniqueName: \"kubernetes.io/projected/e993edd4-f924-4ccc-b818-6718b9c8d7b3-kube-api-access-568gf\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.906549 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b8f843ff2aeeaf7f92dd8ce90b04423c2624f45f2cb345fee3b2851505a7bc2" Nov 24 14:33:30 crc kubenswrapper[4824]: I1124 14:33:30.906669 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-wzj6c" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.022859 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e993edd4-f924-4ccc-b818-6718b9c8d7b3" path="/var/lib/kubelet/pods/e993edd4-f924-4ccc-b818-6718b9c8d7b3/volumes" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.509716 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9hlbw/crc-debug-hfwlv"] Nov 24 14:33:31 crc kubenswrapper[4824]: E1124 14:33:31.510131 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e993edd4-f924-4ccc-b818-6718b9c8d7b3" containerName="container-00" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.510145 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e993edd4-f924-4ccc-b818-6718b9c8d7b3" containerName="container-00" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.510302 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e993edd4-f924-4ccc-b818-6718b9c8d7b3" containerName="container-00" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.510938 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.653364 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl7kd\" (UniqueName: \"kubernetes.io/projected/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-kube-api-access-wl7kd\") pod \"crc-debug-hfwlv\" (UID: \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\") " pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.653475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-host\") pod \"crc-debug-hfwlv\" (UID: \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\") " pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.754752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl7kd\" (UniqueName: \"kubernetes.io/projected/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-kube-api-access-wl7kd\") pod \"crc-debug-hfwlv\" (UID: \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\") " pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.754904 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-host\") pod \"crc-debug-hfwlv\" (UID: \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\") " pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.755024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-host\") pod \"crc-debug-hfwlv\" (UID: \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\") " pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:31 crc kubenswrapper[4824]: I1124 14:33:31.880784 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl7kd\" (UniqueName: \"kubernetes.io/projected/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-kube-api-access-wl7kd\") pod \"crc-debug-hfwlv\" (UID: \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\") " pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:32 crc kubenswrapper[4824]: I1124 14:33:32.131765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:32 crc kubenswrapper[4824]: I1124 14:33:32.930534 4824 generic.go:334] "Generic (PLEG): container finished" podID="8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8" containerID="cab87cc4a44c2a3730d93fec6df63c1c505c84685be08d44e947e8ea36dd6002" exitCode=0 Nov 24 14:33:32 crc kubenswrapper[4824]: I1124 14:33:32.930654 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" event={"ID":"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8","Type":"ContainerDied","Data":"cab87cc4a44c2a3730d93fec6df63c1c505c84685be08d44e947e8ea36dd6002"} Nov 24 14:33:32 crc kubenswrapper[4824]: I1124 14:33:32.931439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" event={"ID":"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8","Type":"ContainerStarted","Data":"536ecada4905dc869f47bf2be9d1b5027df8032a3ffc70ca6fa56ce443dd9496"} Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.088123 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.201163 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-host\") pod \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\" (UID: \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\") " Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.201251 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-host" (OuterVolumeSpecName: "host") pod "8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8" (UID: "8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.202320 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl7kd\" (UniqueName: \"kubernetes.io/projected/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-kube-api-access-wl7kd\") pod \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\" (UID: \"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8\") " Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.203283 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.210380 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-kube-api-access-wl7kd" (OuterVolumeSpecName: "kube-api-access-wl7kd") pod "8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8" (UID: "8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8"). InnerVolumeSpecName "kube-api-access-wl7kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.304433 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl7kd\" (UniqueName: \"kubernetes.io/projected/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8-kube-api-access-wl7kd\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.947007 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" event={"ID":"8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8","Type":"ContainerDied","Data":"536ecada4905dc869f47bf2be9d1b5027df8032a3ffc70ca6fa56ce443dd9496"} Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.947320 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="536ecada4905dc869f47bf2be9d1b5027df8032a3ffc70ca6fa56ce443dd9496" Nov 24 14:33:34 crc kubenswrapper[4824]: I1124 14:33:34.947062 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-hfwlv" Nov 24 14:33:35 crc kubenswrapper[4824]: I1124 14:33:35.068847 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9hlbw/crc-debug-hfwlv"] Nov 24 14:33:35 crc kubenswrapper[4824]: I1124 14:33:35.075645 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9hlbw/crc-debug-hfwlv"] Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.466085 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9hlbw/crc-debug-ttnqx"] Nov 24 14:33:36 crc kubenswrapper[4824]: E1124 14:33:36.466828 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8" containerName="container-00" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.466844 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8" containerName="container-00" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.467059 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8" containerName="container-00" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.467669 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.642823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hscsw\" (UniqueName: \"kubernetes.io/projected/a8db78cc-194c-4451-bbe2-313e63364841-kube-api-access-hscsw\") pod \"crc-debug-ttnqx\" (UID: \"a8db78cc-194c-4451-bbe2-313e63364841\") " pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.643420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a8db78cc-194c-4451-bbe2-313e63364841-host\") pod \"crc-debug-ttnqx\" (UID: \"a8db78cc-194c-4451-bbe2-313e63364841\") " pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.745039 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a8db78cc-194c-4451-bbe2-313e63364841-host\") pod \"crc-debug-ttnqx\" (UID: \"a8db78cc-194c-4451-bbe2-313e63364841\") " pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.745247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hscsw\" (UniqueName: \"kubernetes.io/projected/a8db78cc-194c-4451-bbe2-313e63364841-kube-api-access-hscsw\") pod \"crc-debug-ttnqx\" (UID: \"a8db78cc-194c-4451-bbe2-313e63364841\") " pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.745146 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a8db78cc-194c-4451-bbe2-313e63364841-host\") pod \"crc-debug-ttnqx\" (UID: \"a8db78cc-194c-4451-bbe2-313e63364841\") " pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.762718 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hscsw\" (UniqueName: \"kubernetes.io/projected/a8db78cc-194c-4451-bbe2-313e63364841-kube-api-access-hscsw\") pod \"crc-debug-ttnqx\" (UID: \"a8db78cc-194c-4451-bbe2-313e63364841\") " pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.783190 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:36 crc kubenswrapper[4824]: I1124 14:33:36.963311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" event={"ID":"a8db78cc-194c-4451-bbe2-313e63364841","Type":"ContainerStarted","Data":"05fd0fc8a96afbe81433810a94011a413a6f5010e7105344bbefa5a09f8f6a47"} Nov 24 14:33:37 crc kubenswrapper[4824]: I1124 14:33:37.022608 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8" path="/var/lib/kubelet/pods/8352a1ff-c45a-42e1-b1ee-f05d7b1f94c8/volumes" Nov 24 14:33:37 crc kubenswrapper[4824]: I1124 14:33:37.973648 4824 generic.go:334] "Generic (PLEG): container finished" podID="a8db78cc-194c-4451-bbe2-313e63364841" containerID="da7ac9716795e84c28b453d927c4d70002a8b50438c1a4b3e71123b2ca61aacb" exitCode=0 Nov 24 14:33:37 crc kubenswrapper[4824]: I1124 14:33:37.973706 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" event={"ID":"a8db78cc-194c-4451-bbe2-313e63364841","Type":"ContainerDied","Data":"da7ac9716795e84c28b453d927c4d70002a8b50438c1a4b3e71123b2ca61aacb"} Nov 24 14:33:38 crc kubenswrapper[4824]: I1124 14:33:38.007529 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9hlbw/crc-debug-ttnqx"] Nov 24 14:33:38 crc kubenswrapper[4824]: I1124 14:33:38.015687 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9hlbw/crc-debug-ttnqx"] Nov 24 14:33:39 crc kubenswrapper[4824]: I1124 14:33:39.069879 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:39 crc kubenswrapper[4824]: I1124 14:33:39.228993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hscsw\" (UniqueName: \"kubernetes.io/projected/a8db78cc-194c-4451-bbe2-313e63364841-kube-api-access-hscsw\") pod \"a8db78cc-194c-4451-bbe2-313e63364841\" (UID: \"a8db78cc-194c-4451-bbe2-313e63364841\") " Nov 24 14:33:39 crc kubenswrapper[4824]: I1124 14:33:39.229173 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a8db78cc-194c-4451-bbe2-313e63364841-host\") pod \"a8db78cc-194c-4451-bbe2-313e63364841\" (UID: \"a8db78cc-194c-4451-bbe2-313e63364841\") " Nov 24 14:33:39 crc kubenswrapper[4824]: I1124 14:33:39.229355 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8db78cc-194c-4451-bbe2-313e63364841-host" (OuterVolumeSpecName: "host") pod "a8db78cc-194c-4451-bbe2-313e63364841" (UID: "a8db78cc-194c-4451-bbe2-313e63364841"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:33:39 crc kubenswrapper[4824]: I1124 14:33:39.229781 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a8db78cc-194c-4451-bbe2-313e63364841-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:39 crc kubenswrapper[4824]: I1124 14:33:39.235320 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8db78cc-194c-4451-bbe2-313e63364841-kube-api-access-hscsw" (OuterVolumeSpecName: "kube-api-access-hscsw") pod "a8db78cc-194c-4451-bbe2-313e63364841" (UID: "a8db78cc-194c-4451-bbe2-313e63364841"). InnerVolumeSpecName "kube-api-access-hscsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:33:39 crc kubenswrapper[4824]: I1124 14:33:39.330939 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hscsw\" (UniqueName: \"kubernetes.io/projected/a8db78cc-194c-4451-bbe2-313e63364841-kube-api-access-hscsw\") on node \"crc\" DevicePath \"\"" Nov 24 14:33:39 crc kubenswrapper[4824]: I1124 14:33:39.989763 4824 scope.go:117] "RemoveContainer" containerID="da7ac9716795e84c28b453d927c4d70002a8b50438c1a4b3e71123b2ca61aacb" Nov 24 14:33:39 crc kubenswrapper[4824]: I1124 14:33:39.989794 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/crc-debug-ttnqx" Nov 24 14:33:40 crc kubenswrapper[4824]: I1124 14:33:40.010075 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:33:40 crc kubenswrapper[4824]: E1124 14:33:40.010460 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:33:41 crc kubenswrapper[4824]: I1124 14:33:41.021276 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8db78cc-194c-4451-bbe2-313e63364841" path="/var/lib/kubelet/pods/a8db78cc-194c-4451-bbe2-313e63364841/volumes" Nov 24 14:33:52 crc kubenswrapper[4824]: I1124 14:33:52.010346 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:33:52 crc kubenswrapper[4824]: E1124 14:33:52.011019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:34:04 crc kubenswrapper[4824]: I1124 14:34:04.011454 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:34:04 crc kubenswrapper[4824]: E1124 14:34:04.012047 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.331574 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-78jnn"] Nov 24 14:34:10 crc kubenswrapper[4824]: E1124 14:34:10.332693 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8db78cc-194c-4451-bbe2-313e63364841" containerName="container-00" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.332711 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8db78cc-194c-4451-bbe2-313e63364841" containerName="container-00" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.332975 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8db78cc-194c-4451-bbe2-313e63364841" containerName="container-00" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.334467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.342779 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-78jnn"] Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.445176 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-catalog-content\") pod \"redhat-marketplace-78jnn\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.445331 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpl68\" (UniqueName: \"kubernetes.io/projected/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-kube-api-access-jpl68\") pod \"redhat-marketplace-78jnn\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.445386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-utilities\") pod \"redhat-marketplace-78jnn\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.546555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpl68\" (UniqueName: \"kubernetes.io/projected/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-kube-api-access-jpl68\") pod \"redhat-marketplace-78jnn\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.546626 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-utilities\") pod \"redhat-marketplace-78jnn\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.546697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-catalog-content\") pod \"redhat-marketplace-78jnn\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.547319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-catalog-content\") pod \"redhat-marketplace-78jnn\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.547914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-utilities\") pod \"redhat-marketplace-78jnn\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.568115 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpl68\" (UniqueName: \"kubernetes.io/projected/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-kube-api-access-jpl68\") pod \"redhat-marketplace-78jnn\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:10 crc kubenswrapper[4824]: I1124 14:34:10.660924 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:11 crc kubenswrapper[4824]: I1124 14:34:11.227796 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-78jnn"] Nov 24 14:34:11 crc kubenswrapper[4824]: I1124 14:34:11.312329 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78jnn" event={"ID":"13a3402a-3659-4d2a-93b3-e0fcdb7b273b","Type":"ContainerStarted","Data":"4944430e064cacbf004e1af53a4d734ec7f821454a2781167fe6c9ad23da1dc5"} Nov 24 14:34:12 crc kubenswrapper[4824]: I1124 14:34:12.323029 4824 generic.go:334] "Generic (PLEG): container finished" podID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerID="c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f" exitCode=0 Nov 24 14:34:12 crc kubenswrapper[4824]: I1124 14:34:12.324060 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78jnn" event={"ID":"13a3402a-3659-4d2a-93b3-e0fcdb7b273b","Type":"ContainerDied","Data":"c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f"} Nov 24 14:34:13 crc kubenswrapper[4824]: I1124 14:34:13.333251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78jnn" event={"ID":"13a3402a-3659-4d2a-93b3-e0fcdb7b273b","Type":"ContainerStarted","Data":"c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1"} Nov 24 14:34:14 crc kubenswrapper[4824]: I1124 14:34:14.346466 4824 generic.go:334] "Generic (PLEG): container finished" podID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerID="c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1" exitCode=0 Nov 24 14:34:14 crc kubenswrapper[4824]: I1124 14:34:14.346585 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78jnn" event={"ID":"13a3402a-3659-4d2a-93b3-e0fcdb7b273b","Type":"ContainerDied","Data":"c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1"} Nov 24 14:34:15 crc kubenswrapper[4824]: I1124 14:34:15.356767 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78jnn" event={"ID":"13a3402a-3659-4d2a-93b3-e0fcdb7b273b","Type":"ContainerStarted","Data":"dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133"} Nov 24 14:34:15 crc kubenswrapper[4824]: I1124 14:34:15.391976 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-78jnn" podStartSLOduration=2.729038124 podStartE2EDuration="5.391956904s" podCreationTimestamp="2025-11-24 14:34:10 +0000 UTC" firstStartedPulling="2025-11-24 14:34:12.337397075 +0000 UTC m=+4793.976936385" lastFinishedPulling="2025-11-24 14:34:15.000315845 +0000 UTC m=+4796.639855165" observedRunningTime="2025-11-24 14:34:15.372398837 +0000 UTC m=+4797.011938147" watchObservedRunningTime="2025-11-24 14:34:15.391956904 +0000 UTC m=+4797.031496214" Nov 24 14:34:16 crc kubenswrapper[4824]: I1124 14:34:16.672759 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7545588bcd-5szvs_d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960/barbican-api/0.log" Nov 24 14:34:16 crc kubenswrapper[4824]: I1124 14:34:16.879085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7545588bcd-5szvs_d19d5c0e-cf0b-46a8-b1b7-e832c2fd2960/barbican-api-log/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.000264 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c785d8464-nfqng_6534eeea-5982-4010-a073-f307d7d58036/barbican-keystone-listener/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.013432 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c785d8464-nfqng_6534eeea-5982-4010-a073-f307d7d58036/barbican-keystone-listener-log/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.212789 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b664bf49f-czcvh_b227cafd-bcc4-4e74-96ea-0935efd47fcc/barbican-worker/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.234308 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b664bf49f-czcvh_b227cafd-bcc4-4e74-96ea-0935efd47fcc/barbican-worker-log/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.463394 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rgphs_5ccce3a8-0a97-4a7f-aa2b-0aab163a49a7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.551651 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f521f794-dddf-4c12-9b64-665b541d43ec/ceilometer-notification-agent/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.555647 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f521f794-dddf-4c12-9b64-665b541d43ec/ceilometer-central-agent/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.679669 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f521f794-dddf-4c12-9b64-665b541d43ec/proxy-httpd/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.754332 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f521f794-dddf-4c12-9b64-665b541d43ec/sg-core/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.814048 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4e42a7c2-0db7-4c5d-b731-b543c875ebb0/cinder-api/0.log" Nov 24 14:34:17 crc kubenswrapper[4824]: I1124 14:34:17.933764 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4e42a7c2-0db7-4c5d-b731-b543c875ebb0/cinder-api-log/0.log" Nov 24 14:34:18 crc kubenswrapper[4824]: I1124 14:34:18.010548 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:34:18 crc kubenswrapper[4824]: E1124 14:34:18.010814 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:34:18 crc kubenswrapper[4824]: I1124 14:34:18.092214 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_51aacb45-4a5f-42d3-b905-a051a14856f4/cinder-scheduler/0.log" Nov 24 14:34:18 crc kubenswrapper[4824]: I1124 14:34:18.143479 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_51aacb45-4a5f-42d3-b905-a051a14856f4/probe/0.log" Nov 24 14:34:18 crc kubenswrapper[4824]: I1124 14:34:18.326223 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-k5rds_18493374-ba71-4ac2-97dd-171111137b81/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:18 crc kubenswrapper[4824]: I1124 14:34:18.400156 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tvgk8_92ed5101-5b03-4be3-bd82-32ba9915ba15/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:18 crc kubenswrapper[4824]: I1124 14:34:18.537550 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-hc64g_0bec1542-f209-4da4-a540-9be4708ab637/init/0.log" Nov 24 14:34:18 crc kubenswrapper[4824]: I1124 14:34:18.775332 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-cgxpm_73fc4178-cf60-40cb-b3f0-ea7dc1b1bada/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:18 crc kubenswrapper[4824]: I1124 14:34:18.883823 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-hc64g_0bec1542-f209-4da4-a540-9be4708ab637/init/0.log" Nov 24 14:34:18 crc kubenswrapper[4824]: I1124 14:34:18.930705 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-hc64g_0bec1542-f209-4da4-a540-9be4708ab637/dnsmasq-dns/0.log" Nov 24 14:34:19 crc kubenswrapper[4824]: I1124 14:34:19.116253 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f7321bd0-b6b4-45dc-a4c6-c7a26b601b17/glance-httpd/0.log" Nov 24 14:34:19 crc kubenswrapper[4824]: I1124 14:34:19.258503 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f7321bd0-b6b4-45dc-a4c6-c7a26b601b17/glance-log/0.log" Nov 24 14:34:19 crc kubenswrapper[4824]: I1124 14:34:19.438792 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_dda48846-9717-4dd4-b4af-afe6887923bb/glance-httpd/0.log" Nov 24 14:34:19 crc kubenswrapper[4824]: I1124 14:34:19.451476 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_dda48846-9717-4dd4-b4af-afe6887923bb/glance-log/0.log" Nov 24 14:34:19 crc kubenswrapper[4824]: I1124 14:34:19.622015 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-959bb6dd8-8pcrj_5828a51f-d146-44ea-b8ea-423e017939d8/horizon/1.log" Nov 24 14:34:19 crc kubenswrapper[4824]: I1124 14:34:19.761924 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-959bb6dd8-8pcrj_5828a51f-d146-44ea-b8ea-423e017939d8/horizon/0.log" Nov 24 14:34:19 crc kubenswrapper[4824]: I1124 14:34:19.990768 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-2kqcv_27890aad-2f36-400e-8535-afecafb22a37/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:20 crc kubenswrapper[4824]: I1124 14:34:20.196666 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-959bb6dd8-8pcrj_5828a51f-d146-44ea-b8ea-423e017939d8/horizon-log/0.log" Nov 24 14:34:20 crc kubenswrapper[4824]: I1124 14:34:20.340498 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nhzbn_27ecdc8a-4391-4eb5-a900-0f28984fd1d4/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:20 crc kubenswrapper[4824]: I1124 14:34:20.661792 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:20 crc kubenswrapper[4824]: I1124 14:34:20.662071 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:20 crc kubenswrapper[4824]: I1124 14:34:20.699568 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399881-7g6ln_d24ebeb5-8ce9-42cb-95c6-29162b65eb35/keystone-cron/0.log" Nov 24 14:34:20 crc kubenswrapper[4824]: I1124 14:34:20.724666 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:20 crc kubenswrapper[4824]: I1124 14:34:20.767290 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8a804c88-d42d-43ab-b316-d196ba26f5d1/kube-state-metrics/0.log" Nov 24 14:34:20 crc kubenswrapper[4824]: I1124 14:34:20.937849 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-57d8799bcb-x76nj_378c23b7-2f70-47a0-b605-67087ed59e70/keystone-api/0.log" Nov 24 14:34:21 crc kubenswrapper[4824]: I1124 14:34:21.268172 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-smm76_fa9782de-a426-4c52-a142-335c78e55a33/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:21 crc kubenswrapper[4824]: I1124 14:34:21.475371 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:21 crc kubenswrapper[4824]: I1124 14:34:21.541285 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-78jnn"] Nov 24 14:34:21 crc kubenswrapper[4824]: I1124 14:34:21.913852 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xv9kt_18802c46-869d-4e00-848e-79df8665851d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:22 crc kubenswrapper[4824]: I1124 14:34:22.154937 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-798f5d656f-r9xcp_bf77fc79-5c9e-4688-be73-b390bef539fd/neutron-httpd/0.log" Nov 24 14:34:22 crc kubenswrapper[4824]: I1124 14:34:22.608249 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-798f5d656f-r9xcp_bf77fc79-5c9e-4688-be73-b390bef539fd/neutron-api/0.log" Nov 24 14:34:23 crc kubenswrapper[4824]: I1124 14:34:23.122547 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e5819ad8-0aa5-4d68-831e-09c741267fd5/nova-cell0-conductor-conductor/0.log" Nov 24 14:34:23 crc kubenswrapper[4824]: I1124 14:34:23.323222 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_73aeee43-4360-4af8-9c1e-e1d9bc712157/nova-cell1-conductor-conductor/0.log" Nov 24 14:34:23 crc kubenswrapper[4824]: I1124 14:34:23.432627 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-78jnn" podUID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerName="registry-server" containerID="cri-o://dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133" gracePeriod=2 Nov 24 14:34:23 crc kubenswrapper[4824]: I1124 14:34:23.781373 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ebf76616-c332-4ed1-a188-99843d4692d3/nova-api-log/0.log" Nov 24 14:34:23 crc kubenswrapper[4824]: I1124 14:34:23.904314 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9be10433-b401-4853-ad5a-f32b7668fde2/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 14:34:23 crc kubenswrapper[4824]: I1124 14:34:23.926760 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.012749 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpl68\" (UniqueName: \"kubernetes.io/projected/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-kube-api-access-jpl68\") pod \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.012930 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-catalog-content\") pod \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.013074 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-utilities\") pod \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\" (UID: \"13a3402a-3659-4d2a-93b3-e0fcdb7b273b\") " Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.014578 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-utilities" (OuterVolumeSpecName: "utilities") pod "13a3402a-3659-4d2a-93b3-e0fcdb7b273b" (UID: "13a3402a-3659-4d2a-93b3-e0fcdb7b273b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.031538 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13a3402a-3659-4d2a-93b3-e0fcdb7b273b" (UID: "13a3402a-3659-4d2a-93b3-e0fcdb7b273b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.045070 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-kube-api-access-jpl68" (OuterVolumeSpecName: "kube-api-access-jpl68") pod "13a3402a-3659-4d2a-93b3-e0fcdb7b273b" (UID: "13a3402a-3659-4d2a-93b3-e0fcdb7b273b"). InnerVolumeSpecName "kube-api-access-jpl68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.115228 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.115533 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.115545 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpl68\" (UniqueName: \"kubernetes.io/projected/13a3402a-3659-4d2a-93b3-e0fcdb7b273b-kube-api-access-jpl68\") on node \"crc\" DevicePath \"\"" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.202526 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-qtdx4_4c37785a-1f6f-44c9-8673-b7f9f465682a/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.271040 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8cd73a38-f125-4e8d-b902-f8bc08c1122a/nova-metadata-log/0.log" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.298256 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ebf76616-c332-4ed1-a188-99843d4692d3/nova-api-api/0.log" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.442086 4824 generic.go:334] "Generic (PLEG): container finished" podID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerID="dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133" exitCode=0 Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.442136 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78jnn" event={"ID":"13a3402a-3659-4d2a-93b3-e0fcdb7b273b","Type":"ContainerDied","Data":"dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133"} Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.442161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78jnn" event={"ID":"13a3402a-3659-4d2a-93b3-e0fcdb7b273b","Type":"ContainerDied","Data":"4944430e064cacbf004e1af53a4d734ec7f821454a2781167fe6c9ad23da1dc5"} Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.442178 4824 scope.go:117] "RemoveContainer" containerID="dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.446532 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78jnn" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.466000 4824 scope.go:117] "RemoveContainer" containerID="c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.506256 4824 scope.go:117] "RemoveContainer" containerID="c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.537934 4824 scope.go:117] "RemoveContainer" containerID="dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133" Nov 24 14:34:24 crc kubenswrapper[4824]: E1124 14:34:24.540357 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133\": container with ID starting with dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133 not found: ID does not exist" containerID="dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.540401 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133"} err="failed to get container status \"dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133\": rpc error: code = NotFound desc = could not find container \"dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133\": container with ID starting with dc684d507b9c70bdbfd6e90e3a4c35f08480b5adeefa42d4030a659b93e74133 not found: ID does not exist" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.540424 4824 scope.go:117] "RemoveContainer" containerID="c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1" Nov 24 14:34:24 crc kubenswrapper[4824]: E1124 14:34:24.546255 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1\": container with ID starting with c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1 not found: ID does not exist" containerID="c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.546303 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1"} err="failed to get container status \"c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1\": rpc error: code = NotFound desc = could not find container \"c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1\": container with ID starting with c9ea128c845a6047ee49fb04df38303b957ecbf367201e84e0ddc410b8ac3aa1 not found: ID does not exist" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.546345 4824 scope.go:117] "RemoveContainer" containerID="c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.546445 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-78jnn"] Nov 24 14:34:24 crc kubenswrapper[4824]: E1124 14:34:24.546709 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f\": container with ID starting with c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f not found: ID does not exist" containerID="c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.546730 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f"} err="failed to get container status \"c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f\": rpc error: code = NotFound desc = could not find container \"c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f\": container with ID starting with c1f73a885590f0aebb415f39b5610d201dd31350616344d752948211ab2a9e8f not found: ID does not exist" Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.553806 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-78jnn"] Nov 24 14:34:24 crc kubenswrapper[4824]: I1124 14:34:24.752190 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_90325e3b-c1cb-4244-a81b-fa666652304b/mysql-bootstrap/0.log" Nov 24 14:34:25 crc kubenswrapper[4824]: I1124 14:34:25.028438 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" path="/var/lib/kubelet/pods/13a3402a-3659-4d2a-93b3-e0fcdb7b273b/volumes" Nov 24 14:34:25 crc kubenswrapper[4824]: I1124 14:34:25.064179 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_90325e3b-c1cb-4244-a81b-fa666652304b/galera/0.log" Nov 24 14:34:25 crc kubenswrapper[4824]: I1124 14:34:25.071002 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_90325e3b-c1cb-4244-a81b-fa666652304b/mysql-bootstrap/0.log" Nov 24 14:34:25 crc kubenswrapper[4824]: I1124 14:34:25.184394 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_51c22644-2997-44ff-ab0d-13ce1a067aa2/nova-scheduler-scheduler/0.log" Nov 24 14:34:25 crc kubenswrapper[4824]: I1124 14:34:25.301603 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1e8da990-0e2a-47fd-b7d1-dec77870b40c/mysql-bootstrap/0.log" Nov 24 14:34:25 crc kubenswrapper[4824]: I1124 14:34:25.742122 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1e8da990-0e2a-47fd-b7d1-dec77870b40c/galera/0.log" Nov 24 14:34:25 crc kubenswrapper[4824]: I1124 14:34:25.765274 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1e8da990-0e2a-47fd-b7d1-dec77870b40c/mysql-bootstrap/0.log" Nov 24 14:34:26 crc kubenswrapper[4824]: I1124 14:34:26.009030 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_7d1c1940-0b8b-4b09-ae2a-76168c5a873b/openstackclient/0.log" Nov 24 14:34:26 crc kubenswrapper[4824]: I1124 14:34:26.044568 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ds8fs_b3c188e2-9a89-47fe-8efe-1187a3c81552/ovn-controller/0.log" Nov 24 14:34:26 crc kubenswrapper[4824]: I1124 14:34:26.305209 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-wb2pp_e586afb0-4f13-437a-b00f-ea7e1f6d82ad/openstack-network-exporter/0.log" Nov 24 14:34:26 crc kubenswrapper[4824]: I1124 14:34:26.340673 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8cd73a38-f125-4e8d-b902-f8bc08c1122a/nova-metadata-metadata/0.log" Nov 24 14:34:26 crc kubenswrapper[4824]: I1124 14:34:26.537496 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jvdgl_25645cc4-c970-453f-a6c3-08545ce4c84c/ovsdb-server-init/0.log" Nov 24 14:34:26 crc kubenswrapper[4824]: I1124 14:34:26.752071 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jvdgl_25645cc4-c970-453f-a6c3-08545ce4c84c/ovsdb-server/0.log" Nov 24 14:34:26 crc kubenswrapper[4824]: I1124 14:34:26.777101 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jvdgl_25645cc4-c970-453f-a6c3-08545ce4c84c/ovsdb-server-init/0.log" Nov 24 14:34:26 crc kubenswrapper[4824]: I1124 14:34:26.829849 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jvdgl_25645cc4-c970-453f-a6c3-08545ce4c84c/ovs-vswitchd/0.log" Nov 24 14:34:27 crc kubenswrapper[4824]: I1124 14:34:27.045785 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-bwlv7_f5eaa6e7-da63-4213-ac5d-57fac9f6cd63/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:27 crc kubenswrapper[4824]: I1124 14:34:27.121282 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_102b3dcf-b337-4f91-954b-a4996ba67ca7/openstack-network-exporter/0.log" Nov 24 14:34:27 crc kubenswrapper[4824]: I1124 14:34:27.187067 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_102b3dcf-b337-4f91-954b-a4996ba67ca7/ovn-northd/0.log" Nov 24 14:34:27 crc kubenswrapper[4824]: I1124 14:34:27.328116 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7c045328-9e32-430c-b29b-91a3aeae62c7/ovsdbserver-nb/0.log" Nov 24 14:34:27 crc kubenswrapper[4824]: I1124 14:34:27.347158 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7c045328-9e32-430c-b29b-91a3aeae62c7/openstack-network-exporter/0.log" Nov 24 14:34:27 crc kubenswrapper[4824]: I1124 14:34:27.870524 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c11c797e-a7a6-4b6f-8261-cdb6bdc7a525/ovsdbserver-sb/0.log" Nov 24 14:34:27 crc kubenswrapper[4824]: I1124 14:34:27.933618 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c11c797e-a7a6-4b6f-8261-cdb6bdc7a525/openstack-network-exporter/0.log" Nov 24 14:34:28 crc kubenswrapper[4824]: I1124 14:34:28.140895 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bdb7b6678-pnfdp_d2b3f41b-4bd4-4fcf-b248-6efd3b68f512/placement-api/0.log" Nov 24 14:34:28 crc kubenswrapper[4824]: I1124 14:34:28.220205 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_829719ca-f478-456b-9bbc-07cad8469731/setup-container/0.log" Nov 24 14:34:28 crc kubenswrapper[4824]: I1124 14:34:28.361222 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bdb7b6678-pnfdp_d2b3f41b-4bd4-4fcf-b248-6efd3b68f512/placement-log/0.log" Nov 24 14:34:28 crc kubenswrapper[4824]: I1124 14:34:28.622253 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_829719ca-f478-456b-9bbc-07cad8469731/setup-container/0.log" Nov 24 14:34:28 crc kubenswrapper[4824]: I1124 14:34:28.628770 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_829719ca-f478-456b-9bbc-07cad8469731/rabbitmq/0.log" Nov 24 14:34:28 crc kubenswrapper[4824]: I1124 14:34:28.676147 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98ec6c14-59af-4b38-8e01-89ff157036ca/setup-container/0.log" Nov 24 14:34:28 crc kubenswrapper[4824]: I1124 14:34:28.890539 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98ec6c14-59af-4b38-8e01-89ff157036ca/rabbitmq/0.log" Nov 24 14:34:28 crc kubenswrapper[4824]: I1124 14:34:28.920944 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_98ec6c14-59af-4b38-8e01-89ff157036ca/setup-container/0.log" Nov 24 14:34:28 crc kubenswrapper[4824]: I1124 14:34:28.983769 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-z5v5h_20a6cb74-657c-42ba-a789-411e312d605b/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:29 crc kubenswrapper[4824]: I1124 14:34:29.012427 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:34:29 crc kubenswrapper[4824]: E1124 14:34:29.012756 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:34:29 crc kubenswrapper[4824]: I1124 14:34:29.159352 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-94hhr_53c68f20-6a1f-449e-a54e-aa1df1bf7d32/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:29 crc kubenswrapper[4824]: I1124 14:34:29.593200 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-k7zrh_1a2dbe7f-7854-4dc7-9511-bba8a8c81a08/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:29 crc kubenswrapper[4824]: I1124 14:34:29.699049 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-97k6l_9ed4df3b-472c-4815-a5ef-66125b2dde81/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:29 crc kubenswrapper[4824]: I1124 14:34:29.905524 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-vtnl7_62b46948-d9cc-4a68-8d94-70a7538bd61d/ssh-known-hosts-edpm-deployment/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.140632 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8cc59fd67-2j94h_57b08eb1-c1bf-4b56-880f-4d8adeaa32fc/proxy-httpd/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.252607 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8cc59fd67-2j94h_57b08eb1-c1bf-4b56-880f-4d8adeaa32fc/proxy-server/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.354537 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-m5pdt_c2abd860-636d-4e37-8a85-334aefcf44ec/swift-ring-rebalance/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.525179 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/account-auditor/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.579889 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/account-reaper/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.700382 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/account-replicator/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.768296 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/account-server/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.794526 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/container-auditor/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.840537 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/container-replicator/0.log" Nov 24 14:34:30 crc kubenswrapper[4824]: I1124 14:34:30.945380 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/container-server/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.061730 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-auditor/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.069549 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-expirer/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.102968 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/container-updater/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.243809 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-replicator/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.294869 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-updater/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.361504 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/object-server/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.367180 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/rsync/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.598843 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1ac7bb70-1d1a-4ef5-b5f2-134792dcec5c/swift-recon-cron/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.757901 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-k6x98_3fdb12da-712c-4ebe-97d1-4ebc46a1db80/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:31 crc kubenswrapper[4824]: I1124 14:34:31.888076 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_a95b11ef-b7cf-4ba5-9aef-6ed7b7f035b0/tempest-tests-tempest-tests-runner/0.log" Nov 24 14:34:32 crc kubenswrapper[4824]: I1124 14:34:32.130606 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b719b5e7-33d8-43ec-9d61-19bd0676886a/test-operator-logs-container/0.log" Nov 24 14:34:32 crc kubenswrapper[4824]: I1124 14:34:32.304257 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-j8h8s_ddb976c7-8261-4a6b-96a1-75a8e4b5d282/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:34:43 crc kubenswrapper[4824]: I1124 14:34:43.720503 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_a7687928-0bb1-456c-8555-e918db203197/memcached/0.log" Nov 24 14:34:44 crc kubenswrapper[4824]: I1124 14:34:44.010188 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:34:44 crc kubenswrapper[4824]: E1124 14:34:44.010783 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:34:58 crc kubenswrapper[4824]: I1124 14:34:58.010183 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:34:58 crc kubenswrapper[4824]: E1124 14:34:58.011015 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:35:05 crc kubenswrapper[4824]: I1124 14:35:05.264588 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-747cg_dc25960c-da62-4fe3-97c1-cf8526bc9c1d/kube-rbac-proxy/0.log" Nov 24 14:35:05 crc kubenswrapper[4824]: I1124 14:35:05.362224 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-747cg_dc25960c-da62-4fe3-97c1-cf8526bc9c1d/manager/0.log" Nov 24 14:35:05 crc kubenswrapper[4824]: I1124 14:35:05.527183 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-h5mx8_88cfde0a-21e8-45ca-a7b3-43699c2ff345/kube-rbac-proxy/0.log" Nov 24 14:35:05 crc kubenswrapper[4824]: I1124 14:35:05.703285 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-h5mx8_88cfde0a-21e8-45ca-a7b3-43699c2ff345/manager/0.log" Nov 24 14:35:05 crc kubenswrapper[4824]: I1124 14:35:05.998771 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-gw957_3b95ed18-ec44-4712-9f01-adeef55deb75/kube-rbac-proxy/0.log" Nov 24 14:35:06 crc kubenswrapper[4824]: I1124 14:35:06.044734 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-gw957_3b95ed18-ec44-4712-9f01-adeef55deb75/manager/0.log" Nov 24 14:35:06 crc kubenswrapper[4824]: I1124 14:35:06.460846 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/util/0.log" Nov 24 14:35:06 crc kubenswrapper[4824]: I1124 14:35:06.625150 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/util/0.log" Nov 24 14:35:06 crc kubenswrapper[4824]: I1124 14:35:06.641441 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/pull/0.log" Nov 24 14:35:06 crc kubenswrapper[4824]: I1124 14:35:06.708526 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/pull/0.log" Nov 24 14:35:06 crc kubenswrapper[4824]: I1124 14:35:06.918816 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/pull/0.log" Nov 24 14:35:06 crc kubenswrapper[4824]: I1124 14:35:06.950288 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/extract/0.log" Nov 24 14:35:06 crc kubenswrapper[4824]: I1124 14:35:06.951273 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f435090ed9bde41dbac4461416e374954ebd31c691df9f8d376db660522c57z_d720b626-1d35-4d40-95be-dbd95dcafc27/util/0.log" Nov 24 14:35:07 crc kubenswrapper[4824]: I1124 14:35:07.214132 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-gvhjd_9c973000-24b9-4087-a3eb-c5f087a50ae3/manager/0.log" Nov 24 14:35:07 crc kubenswrapper[4824]: I1124 14:35:07.236473 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-s2k8r_0cac78b6-e909-4616-bce8-d3ce8ac90452/kube-rbac-proxy/0.log" Nov 24 14:35:07 crc kubenswrapper[4824]: I1124 14:35:07.265905 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-gvhjd_9c973000-24b9-4087-a3eb-c5f087a50ae3/kube-rbac-proxy/0.log" Nov 24 14:35:07 crc kubenswrapper[4824]: I1124 14:35:07.463090 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-s2k8r_0cac78b6-e909-4616-bce8-d3ce8ac90452/manager/0.log" Nov 24 14:35:07 crc kubenswrapper[4824]: I1124 14:35:07.465759 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-cdnsz_5f502855-0eb1-455d-8e3b-4af6ac353aa9/kube-rbac-proxy/0.log" Nov 24 14:35:07 crc kubenswrapper[4824]: I1124 14:35:07.558028 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-cdnsz_5f502855-0eb1-455d-8e3b-4af6ac353aa9/manager/0.log" Nov 24 14:35:07 crc kubenswrapper[4824]: I1124 14:35:07.749080 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-rn29q_ec73cd81-275a-44e0-93dc-6b20123a88d3/kube-rbac-proxy/0.log" Nov 24 14:35:07 crc kubenswrapper[4824]: I1124 14:35:07.786081 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-rn29q_ec73cd81-275a-44e0-93dc-6b20123a88d3/manager/0.log" Nov 24 14:35:07 crc kubenswrapper[4824]: I1124 14:35:07.949859 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-cmmdj_05a9eb12-a2a1-4fbe-901b-cb817e4713b4/kube-rbac-proxy/0.log" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.036091 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-cmmdj_05a9eb12-a2a1-4fbe-901b-cb817e4713b4/manager/0.log" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.128351 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-v6dbs_65d269bf-f6af-4239-8e51-bf44fe7830d1/kube-rbac-proxy/0.log" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.239048 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-v6dbs_65d269bf-f6af-4239-8e51-bf44fe7830d1/manager/0.log" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.297370 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-xdrwd_3b3da19f-12b1-413d-b6aa-38bc498172c1/kube-rbac-proxy/0.log" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.351356 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-xdrwd_3b3da19f-12b1-413d-b6aa-38bc498172c1/manager/0.log" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.470162 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tsctv"] Nov 24 14:35:08 crc kubenswrapper[4824]: E1124 14:35:08.470600 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerName="extract-content" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.470618 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerName="extract-content" Nov 24 14:35:08 crc kubenswrapper[4824]: E1124 14:35:08.470630 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerName="registry-server" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.470638 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerName="registry-server" Nov 24 14:35:08 crc kubenswrapper[4824]: E1124 14:35:08.470680 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerName="extract-utilities" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.470689 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerName="extract-utilities" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.470960 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13a3402a-3659-4d2a-93b3-e0fcdb7b273b" containerName="registry-server" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.472719 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.491393 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tsctv"] Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.564019 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jxz8\" (UniqueName: \"kubernetes.io/projected/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-kube-api-access-6jxz8\") pod \"community-operators-tsctv\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.564331 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-catalog-content\") pod \"community-operators-tsctv\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.564482 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-utilities\") pod \"community-operators-tsctv\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.633112 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-mzs9m_9971e72a-bb18-4951-a752-ee96f1725801/kube-rbac-proxy/0.log" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.665633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-catalog-content\") pod \"community-operators-tsctv\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.665710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-utilities\") pod \"community-operators-tsctv\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.665753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jxz8\" (UniqueName: \"kubernetes.io/projected/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-kube-api-access-6jxz8\") pod \"community-operators-tsctv\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.666514 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-catalog-content\") pod \"community-operators-tsctv\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.666737 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-utilities\") pod \"community-operators-tsctv\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.667024 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-mzs9m_9971e72a-bb18-4951-a752-ee96f1725801/manager/0.log" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.696790 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jxz8\" (UniqueName: \"kubernetes.io/projected/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-kube-api-access-6jxz8\") pod \"community-operators-tsctv\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:08 crc kubenswrapper[4824]: I1124 14:35:08.796516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.226550 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-zncj6_50437b59-29ab-4e49-a751-80a83d9e2aa0/kube-rbac-proxy/0.log" Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.274673 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-zncj6_50437b59-29ab-4e49-a751-80a83d9e2aa0/manager/0.log" Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.366011 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tsctv"] Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.491028 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-qzxtx_4cca949a-2334-4d41-bb9e-13e6bd89d7a8/manager/0.log" Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.518227 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-qzxtx_4cca949a-2334-4d41-bb9e-13e6bd89d7a8/kube-rbac-proxy/0.log" Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.567610 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-bg2m4_218ca72b-e49f-4830-8be8-9b6e85956009/kube-rbac-proxy/0.log" Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.911735 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp_99379416-bd58-4014-a297-7b14b8ff4aeb/manager/0.log" Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.915513 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-bg2m4_218ca72b-e49f-4830-8be8-9b6e85956009/manager/0.log" Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.946796 4824 generic.go:334] "Generic (PLEG): container finished" podID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerID="61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051" exitCode=0 Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.947440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tsctv" event={"ID":"143eb412-3120-4a23-b4cd-b2d4a07d3e2a","Type":"ContainerDied","Data":"61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051"} Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.947508 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tsctv" event={"ID":"143eb412-3120-4a23-b4cd-b2d4a07d3e2a","Type":"ContainerStarted","Data":"be2212780368a27bb8250699152faf90f2de1f3a6d8514eb40331eb82090022f"} Nov 24 14:35:09 crc kubenswrapper[4824]: I1124 14:35:09.949486 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:35:10 crc kubenswrapper[4824]: I1124 14:35:10.051202 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-mr8fp_99379416-bd58-4014-a297-7b14b8ff4aeb/kube-rbac-proxy/0.log" Nov 24 14:35:10 crc kubenswrapper[4824]: I1124 14:35:10.219753 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7547bd9b9-tpbcz_4f304aa2-86ae-45a3-9039-848548cbd500/kube-rbac-proxy/0.log" Nov 24 14:35:10 crc kubenswrapper[4824]: I1124 14:35:10.332172 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-866b74d8b7-d68dq_af6dc481-60ca-48ae-a75c-edec34270b7f/kube-rbac-proxy/0.log" Nov 24 14:35:10 crc kubenswrapper[4824]: I1124 14:35:10.552139 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-866b74d8b7-d68dq_af6dc481-60ca-48ae-a75c-edec34270b7f/operator/0.log" Nov 24 14:35:10 crc kubenswrapper[4824]: I1124 14:35:10.610315 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hht55_eaf96b33-49f7-4e4f-86e2-7a8a8b479bba/registry-server/0.log" Nov 24 14:35:10 crc kubenswrapper[4824]: I1124 14:35:10.807343 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-2jjxl_926db806-d562-44f8-9ac9-71f337dbb202/kube-rbac-proxy/0.log" Nov 24 14:35:10 crc kubenswrapper[4824]: I1124 14:35:10.964778 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tsctv" event={"ID":"143eb412-3120-4a23-b4cd-b2d4a07d3e2a","Type":"ContainerStarted","Data":"11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c"} Nov 24 14:35:10 crc kubenswrapper[4824]: I1124 14:35:10.987437 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-rzg64_4676e827-15a7-4286-a4a8-0147b70e315b/kube-rbac-proxy/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.016120 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:35:11 crc kubenswrapper[4824]: E1124 14:35:11.023823 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.035555 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-2jjxl_926db806-d562-44f8-9ac9-71f337dbb202/manager/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.206070 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-rzg64_4676e827-15a7-4286-a4a8-0147b70e315b/manager/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.327430 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-b82s2_c97af640-fa25-46a6-816a-104506790d6d/operator/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.463422 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7547bd9b9-tpbcz_4f304aa2-86ae-45a3-9039-848548cbd500/manager/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.504842 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-hnczv_39f22836-fab7-4ffa-b95f-d2dad15a98b9/kube-rbac-proxy/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.554263 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-hnczv_39f22836-fab7-4ffa-b95f-d2dad15a98b9/manager/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.618795 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-w82fg_3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2/kube-rbac-proxy/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.793993 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-fhn99_bc237c32-95f2-4386-88f3-a337d2269261/kube-rbac-proxy/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.820641 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-w82fg_3f5bebf7-9b0f-480a-a26d-e5d6f7813ef2/manager/0.log" Nov 24 14:35:11 crc kubenswrapper[4824]: I1124 14:35:11.878870 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-fhn99_bc237c32-95f2-4386-88f3-a337d2269261/manager/0.log" Nov 24 14:35:12 crc kubenswrapper[4824]: I1124 14:35:12.004227 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-t46st_c8001b29-70e7-4861-9b3d-41e732db7bf9/kube-rbac-proxy/0.log" Nov 24 14:35:12 crc kubenswrapper[4824]: I1124 14:35:12.066782 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-t46st_c8001b29-70e7-4861-9b3d-41e732db7bf9/manager/0.log" Nov 24 14:35:12 crc kubenswrapper[4824]: I1124 14:35:12.981604 4824 generic.go:334] "Generic (PLEG): container finished" podID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerID="11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c" exitCode=0 Nov 24 14:35:12 crc kubenswrapper[4824]: I1124 14:35:12.981649 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tsctv" event={"ID":"143eb412-3120-4a23-b4cd-b2d4a07d3e2a","Type":"ContainerDied","Data":"11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c"} Nov 24 14:35:13 crc kubenswrapper[4824]: I1124 14:35:13.992295 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tsctv" event={"ID":"143eb412-3120-4a23-b4cd-b2d4a07d3e2a","Type":"ContainerStarted","Data":"4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d"} Nov 24 14:35:14 crc kubenswrapper[4824]: I1124 14:35:14.026756 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tsctv" podStartSLOduration=2.593068941 podStartE2EDuration="6.026738119s" podCreationTimestamp="2025-11-24 14:35:08 +0000 UTC" firstStartedPulling="2025-11-24 14:35:09.949214433 +0000 UTC m=+4851.588753743" lastFinishedPulling="2025-11-24 14:35:13.382883611 +0000 UTC m=+4855.022422921" observedRunningTime="2025-11-24 14:35:14.026104973 +0000 UTC m=+4855.665644283" watchObservedRunningTime="2025-11-24 14:35:14.026738119 +0000 UTC m=+4855.666277429" Nov 24 14:35:18 crc kubenswrapper[4824]: I1124 14:35:18.797201 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:18 crc kubenswrapper[4824]: I1124 14:35:18.797788 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:18 crc kubenswrapper[4824]: I1124 14:35:18.843233 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:19 crc kubenswrapper[4824]: I1124 14:35:19.099353 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:19 crc kubenswrapper[4824]: I1124 14:35:19.155560 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tsctv"] Nov 24 14:35:21 crc kubenswrapper[4824]: I1124 14:35:21.063438 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tsctv" podUID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerName="registry-server" containerID="cri-o://4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d" gracePeriod=2 Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.043085 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.074873 4824 generic.go:334] "Generic (PLEG): container finished" podID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerID="4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d" exitCode=0 Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.074918 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tsctv" event={"ID":"143eb412-3120-4a23-b4cd-b2d4a07d3e2a","Type":"ContainerDied","Data":"4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d"} Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.074948 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tsctv" event={"ID":"143eb412-3120-4a23-b4cd-b2d4a07d3e2a","Type":"ContainerDied","Data":"be2212780368a27bb8250699152faf90f2de1f3a6d8514eb40331eb82090022f"} Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.074953 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tsctv" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.074964 4824 scope.go:117] "RemoveContainer" containerID="4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.098049 4824 scope.go:117] "RemoveContainer" containerID="11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.109148 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jxz8\" (UniqueName: \"kubernetes.io/projected/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-kube-api-access-6jxz8\") pod \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.110726 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-catalog-content\") pod \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.110893 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-utilities\") pod \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\" (UID: \"143eb412-3120-4a23-b4cd-b2d4a07d3e2a\") " Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.115991 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-utilities" (OuterVolumeSpecName: "utilities") pod "143eb412-3120-4a23-b4cd-b2d4a07d3e2a" (UID: "143eb412-3120-4a23-b4cd-b2d4a07d3e2a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.134177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-kube-api-access-6jxz8" (OuterVolumeSpecName: "kube-api-access-6jxz8") pod "143eb412-3120-4a23-b4cd-b2d4a07d3e2a" (UID: "143eb412-3120-4a23-b4cd-b2d4a07d3e2a"). InnerVolumeSpecName "kube-api-access-6jxz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.173396 4824 scope.go:117] "RemoveContainer" containerID="61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.187720 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "143eb412-3120-4a23-b4cd-b2d4a07d3e2a" (UID: "143eb412-3120-4a23-b4cd-b2d4a07d3e2a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.213415 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jxz8\" (UniqueName: \"kubernetes.io/projected/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-kube-api-access-6jxz8\") on node \"crc\" DevicePath \"\"" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.213449 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.213461 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143eb412-3120-4a23-b4cd-b2d4a07d3e2a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.214480 4824 scope.go:117] "RemoveContainer" containerID="4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d" Nov 24 14:35:22 crc kubenswrapper[4824]: E1124 14:35:22.215009 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d\": container with ID starting with 4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d not found: ID does not exist" containerID="4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.215122 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d"} err="failed to get container status \"4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d\": rpc error: code = NotFound desc = could not find container \"4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d\": container with ID starting with 4597ce324a47fa254282aa18b5e8dbf854366063deb5eac8fe393dcdaf3a1a9d not found: ID does not exist" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.215237 4824 scope.go:117] "RemoveContainer" containerID="11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c" Nov 24 14:35:22 crc kubenswrapper[4824]: E1124 14:35:22.215700 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c\": container with ID starting with 11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c not found: ID does not exist" containerID="11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.215830 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c"} err="failed to get container status \"11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c\": rpc error: code = NotFound desc = could not find container \"11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c\": container with ID starting with 11f8ff60fc848ad291fc258d1e7999d515be24544ce852a1699c719d0b86c53c not found: ID does not exist" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.215948 4824 scope.go:117] "RemoveContainer" containerID="61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051" Nov 24 14:35:22 crc kubenswrapper[4824]: E1124 14:35:22.216476 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051\": container with ID starting with 61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051 not found: ID does not exist" containerID="61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.216582 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051"} err="failed to get container status \"61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051\": rpc error: code = NotFound desc = could not find container \"61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051\": container with ID starting with 61c214b356b967e6840175eefd200e59e1218b57425cd3498f68b1173988f051 not found: ID does not exist" Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.428761 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tsctv"] Nov 24 14:35:22 crc kubenswrapper[4824]: I1124 14:35:22.435592 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tsctv"] Nov 24 14:35:23 crc kubenswrapper[4824]: I1124 14:35:23.020484 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" path="/var/lib/kubelet/pods/143eb412-3120-4a23-b4cd-b2d4a07d3e2a/volumes" Nov 24 14:35:27 crc kubenswrapper[4824]: I1124 14:35:27.010054 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:35:27 crc kubenswrapper[4824]: E1124 14:35:27.010522 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-74f4n_openshift-machine-config-operator(b90f50d7-5574-472e-9532-91b491a0cc56)\"" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" Nov 24 14:35:31 crc kubenswrapper[4824]: I1124 14:35:31.069550 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-8h66q_91b65029-b851-4317-a1fc-26b30a5c8734/control-plane-machine-set-operator/0.log" Nov 24 14:35:31 crc kubenswrapper[4824]: I1124 14:35:31.271275 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5r76_8a59be8e-4194-4e1c-89bd-3136366ed13b/machine-api-operator/0.log" Nov 24 14:35:31 crc kubenswrapper[4824]: I1124 14:35:31.312164 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5r76_8a59be8e-4194-4e1c-89bd-3136366ed13b/kube-rbac-proxy/0.log" Nov 24 14:35:41 crc kubenswrapper[4824]: I1124 14:35:41.010407 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:35:41 crc kubenswrapper[4824]: I1124 14:35:41.280800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"852247ad8ec5e01521c52f7bcbd76d7dbfe197d123f160f2a086bf954fa38a02"} Nov 24 14:35:43 crc kubenswrapper[4824]: I1124 14:35:43.749630 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-8mhb8_121b83e4-cc24-4e97-831b-c05b2dce0aca/cert-manager-controller/0.log" Nov 24 14:35:43 crc kubenswrapper[4824]: I1124 14:35:43.904271 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-p8wjb_c65d2285-3c4b-407b-937c-53a7d1302a8a/cert-manager-webhook/0.log" Nov 24 14:35:43 crc kubenswrapper[4824]: I1124 14:35:43.933161 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-4hmfd_db3d2465-e3f8-434a-a279-9de75d0f2bbc/cert-manager-cainjector/0.log" Nov 24 14:35:56 crc kubenswrapper[4824]: I1124 14:35:56.774527 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-7sv5s_a252dabf-79e7-442e-990c-3f15bc47d536/nmstate-console-plugin/0.log" Nov 24 14:35:57 crc kubenswrapper[4824]: I1124 14:35:57.012440 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gw4fj_1e8318a0-5501-47b7-836a-4d650de94219/nmstate-handler/0.log" Nov 24 14:35:57 crc kubenswrapper[4824]: I1124 14:35:57.126567 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-gb5z5_2717ebec-b84b-4d07-a1ac-67209e7e5e32/nmstate-metrics/0.log" Nov 24 14:35:57 crc kubenswrapper[4824]: I1124 14:35:57.143509 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-gb5z5_2717ebec-b84b-4d07-a1ac-67209e7e5e32/kube-rbac-proxy/0.log" Nov 24 14:35:57 crc kubenswrapper[4824]: I1124 14:35:57.412379 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-ghp4s_04d4e05b-9e17-473f-8d25-933b9a4ed60a/nmstate-operator/0.log" Nov 24 14:35:57 crc kubenswrapper[4824]: I1124 14:35:57.440101 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-qs6l4_8876f622-9722-4257-8526-9f1f0e6a3daa/nmstate-webhook/0.log" Nov 24 14:36:13 crc kubenswrapper[4824]: I1124 14:36:13.512660 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-m4d79_077fe5ab-ea92-4aa8-8b76-dc791ae6d41b/kube-rbac-proxy/0.log" Nov 24 14:36:13 crc kubenswrapper[4824]: I1124 14:36:13.704260 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-m4d79_077fe5ab-ea92-4aa8-8b76-dc791ae6d41b/controller/0.log" Nov 24 14:36:13 crc kubenswrapper[4824]: I1124 14:36:13.770905 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-frr-files/0.log" Nov 24 14:36:13 crc kubenswrapper[4824]: I1124 14:36:13.979510 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-reloader/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.019488 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-metrics/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.044344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-frr-files/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.053378 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-reloader/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.246154 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-metrics/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.246202 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-frr-files/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.267272 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-reloader/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.315376 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-metrics/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.466957 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-metrics/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.480037 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-frr-files/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.484208 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/cp-reloader/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.579183 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/controller/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.744150 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/frr-metrics/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.773750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/kube-rbac-proxy/0.log" Nov 24 14:36:14 crc kubenswrapper[4824]: I1124 14:36:14.848872 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/kube-rbac-proxy-frr/0.log" Nov 24 14:36:15 crc kubenswrapper[4824]: I1124 14:36:15.084036 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-2gp4q_db503b31-f22b-4b5c-a20f-970c15435851/frr-k8s-webhook-server/0.log" Nov 24 14:36:15 crc kubenswrapper[4824]: I1124 14:36:15.130597 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/reloader/0.log" Nov 24 14:36:15 crc kubenswrapper[4824]: I1124 14:36:15.462769 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-b6cff747d-x5ng4_99485513-6292-40ee-aede-2ef1d8566124/manager/0.log" Nov 24 14:36:15 crc kubenswrapper[4824]: I1124 14:36:15.611354 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-ffbf5bdd6-fr567_07d6946c-c191-4a18-9903-194bdc00b4a0/webhook-server/0.log" Nov 24 14:36:15 crc kubenswrapper[4824]: I1124 14:36:15.775405 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r5w24_e2bbe7bd-470b-4363-a85c-7a2084da3e4e/kube-rbac-proxy/0.log" Nov 24 14:36:16 crc kubenswrapper[4824]: I1124 14:36:16.169386 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dxtg8_d7ab142b-a25e-4591-9672-6209bf35a3a8/frr/0.log" Nov 24 14:36:16 crc kubenswrapper[4824]: I1124 14:36:16.394049 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r5w24_e2bbe7bd-470b-4363-a85c-7a2084da3e4e/speaker/0.log" Nov 24 14:36:28 crc kubenswrapper[4824]: I1124 14:36:28.989025 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/util/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.162702 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/pull/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.182232 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/pull/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.250326 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/util/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.396277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/util/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.469874 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/pull/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.512071 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edh67s_44dcd7ab-51b6-4472-a91c-940e30d4cac9/extract/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.642385 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-utilities/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.918900 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-content/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.926723 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-utilities/0.log" Nov 24 14:36:29 crc kubenswrapper[4824]: I1124 14:36:29.972253 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-content/0.log" Nov 24 14:36:30 crc kubenswrapper[4824]: I1124 14:36:30.138202 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-content/0.log" Nov 24 14:36:30 crc kubenswrapper[4824]: I1124 14:36:30.271056 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/extract-utilities/0.log" Nov 24 14:36:30 crc kubenswrapper[4824]: I1124 14:36:30.515891 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-utilities/0.log" Nov 24 14:36:30 crc kubenswrapper[4824]: I1124 14:36:30.741251 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-utilities/0.log" Nov 24 14:36:30 crc kubenswrapper[4824]: I1124 14:36:30.792430 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-content/0.log" Nov 24 14:36:30 crc kubenswrapper[4824]: I1124 14:36:30.806624 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-content/0.log" Nov 24 14:36:30 crc kubenswrapper[4824]: I1124 14:36:30.839650 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rqjfm_e2bc66fe-f9be-4fdf-8360-62e62c629cae/registry-server/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.103531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-content/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.124859 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/extract-utilities/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.392327 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/util/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.434827 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nh4bn_fe27b019-f6cf-4eec-83df-ef23b6483cf8/registry-server/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.582672 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/pull/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.607427 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/util/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.618703 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/pull/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.862175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/extract/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.874691 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/pull/0.log" Nov 24 14:36:31 crc kubenswrapper[4824]: I1124 14:36:31.921927 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6f9z8v_0dbae9a3-1221-48ad-84a8-327d9f8176e7/util/0.log" Nov 24 14:36:32 crc kubenswrapper[4824]: I1124 14:36:32.088106 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qkj9f_7b3b4423-493d-4cc4-9533-d87011248427/marketplace-operator/0.log" Nov 24 14:36:32 crc kubenswrapper[4824]: I1124 14:36:32.126454 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-utilities/0.log" Nov 24 14:36:32 crc kubenswrapper[4824]: I1124 14:36:32.392214 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-content/0.log" Nov 24 14:36:32 crc kubenswrapper[4824]: I1124 14:36:32.392333 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-utilities/0.log" Nov 24 14:36:32 crc kubenswrapper[4824]: I1124 14:36:32.453481 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-content/0.log" Nov 24 14:36:32 crc kubenswrapper[4824]: I1124 14:36:32.637390 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-content/0.log" Nov 24 14:36:32 crc kubenswrapper[4824]: I1124 14:36:32.643002 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/extract-utilities/0.log" Nov 24 14:36:32 crc kubenswrapper[4824]: I1124 14:36:32.899090 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dzj4v_4adc5ce2-f4df-4893-9d1b-9bc4c7267faa/registry-server/0.log" Nov 24 14:36:32 crc kubenswrapper[4824]: I1124 14:36:32.955613 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xm5c6_9f0b9253-e525-4c45-9f90-036eceb195d1/extract-utilities/0.log" Nov 24 14:36:33 crc kubenswrapper[4824]: I1124 14:36:33.154280 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xm5c6_9f0b9253-e525-4c45-9f90-036eceb195d1/extract-utilities/0.log" Nov 24 14:36:33 crc kubenswrapper[4824]: I1124 14:36:33.199331 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xm5c6_9f0b9253-e525-4c45-9f90-036eceb195d1/extract-content/0.log" Nov 24 14:36:33 crc kubenswrapper[4824]: I1124 14:36:33.202478 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xm5c6_9f0b9253-e525-4c45-9f90-036eceb195d1/extract-content/0.log" Nov 24 14:36:33 crc kubenswrapper[4824]: I1124 14:36:33.333977 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xm5c6_9f0b9253-e525-4c45-9f90-036eceb195d1/extract-utilities/0.log" Nov 24 14:36:33 crc kubenswrapper[4824]: I1124 14:36:33.344227 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xm5c6_9f0b9253-e525-4c45-9f90-036eceb195d1/extract-content/0.log" Nov 24 14:36:33 crc kubenswrapper[4824]: I1124 14:36:33.492986 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xm5c6_9f0b9253-e525-4c45-9f90-036eceb195d1/registry-server/0.log" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.712496 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5bjxl"] Nov 24 14:37:09 crc kubenswrapper[4824]: E1124 14:37:09.713658 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerName="registry-server" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.713679 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerName="registry-server" Nov 24 14:37:09 crc kubenswrapper[4824]: E1124 14:37:09.713702 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerName="extract-content" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.713711 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerName="extract-content" Nov 24 14:37:09 crc kubenswrapper[4824]: E1124 14:37:09.713747 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerName="extract-utilities" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.713755 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerName="extract-utilities" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.714026 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="143eb412-3120-4a23-b4cd-b2d4a07d3e2a" containerName="registry-server" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.715703 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.732975 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5bjxl"] Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.819443 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-catalog-content\") pod \"certified-operators-5bjxl\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.819529 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8flt8\" (UniqueName: \"kubernetes.io/projected/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-kube-api-access-8flt8\") pod \"certified-operators-5bjxl\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.819593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-utilities\") pod \"certified-operators-5bjxl\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.921152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-utilities\") pod \"certified-operators-5bjxl\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.921356 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-catalog-content\") pod \"certified-operators-5bjxl\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.921417 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8flt8\" (UniqueName: \"kubernetes.io/projected/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-kube-api-access-8flt8\") pod \"certified-operators-5bjxl\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.922250 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-utilities\") pod \"certified-operators-5bjxl\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.922547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-catalog-content\") pod \"certified-operators-5bjxl\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:09 crc kubenswrapper[4824]: I1124 14:37:09.942961 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8flt8\" (UniqueName: \"kubernetes.io/projected/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-kube-api-access-8flt8\") pod \"certified-operators-5bjxl\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:10 crc kubenswrapper[4824]: I1124 14:37:10.036045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:10 crc kubenswrapper[4824]: I1124 14:37:10.950954 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5bjxl"] Nov 24 14:37:11 crc kubenswrapper[4824]: I1124 14:37:11.364853 4824 generic.go:334] "Generic (PLEG): container finished" podID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerID="9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a" exitCode=0 Nov 24 14:37:11 crc kubenswrapper[4824]: I1124 14:37:11.364908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bjxl" event={"ID":"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739","Type":"ContainerDied","Data":"9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a"} Nov 24 14:37:11 crc kubenswrapper[4824]: I1124 14:37:11.365179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bjxl" event={"ID":"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739","Type":"ContainerStarted","Data":"ab0c6687a1c0f82a7968e0541d7d4185373683d62c51ac333e436f955a43ccfe"} Nov 24 14:37:13 crc kubenswrapper[4824]: I1124 14:37:13.389663 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bjxl" event={"ID":"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739","Type":"ContainerStarted","Data":"754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a"} Nov 24 14:37:14 crc kubenswrapper[4824]: I1124 14:37:14.398919 4824 generic.go:334] "Generic (PLEG): container finished" podID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerID="754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a" exitCode=0 Nov 24 14:37:14 crc kubenswrapper[4824]: I1124 14:37:14.398958 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bjxl" event={"ID":"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739","Type":"ContainerDied","Data":"754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a"} Nov 24 14:37:15 crc kubenswrapper[4824]: I1124 14:37:15.409302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bjxl" event={"ID":"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739","Type":"ContainerStarted","Data":"37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b"} Nov 24 14:37:15 crc kubenswrapper[4824]: I1124 14:37:15.437151 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5bjxl" podStartSLOduration=2.973542372 podStartE2EDuration="6.437129314s" podCreationTimestamp="2025-11-24 14:37:09 +0000 UTC" firstStartedPulling="2025-11-24 14:37:11.366425048 +0000 UTC m=+4973.005964358" lastFinishedPulling="2025-11-24 14:37:14.83001199 +0000 UTC m=+4976.469551300" observedRunningTime="2025-11-24 14:37:15.428048278 +0000 UTC m=+4977.067587588" watchObservedRunningTime="2025-11-24 14:37:15.437129314 +0000 UTC m=+4977.076668624" Nov 24 14:37:20 crc kubenswrapper[4824]: I1124 14:37:20.037114 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:20 crc kubenswrapper[4824]: I1124 14:37:20.037671 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:20 crc kubenswrapper[4824]: I1124 14:37:20.084672 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:20 crc kubenswrapper[4824]: I1124 14:37:20.528010 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:21 crc kubenswrapper[4824]: I1124 14:37:21.335361 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5bjxl"] Nov 24 14:37:22 crc kubenswrapper[4824]: I1124 14:37:22.467958 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5bjxl" podUID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerName="registry-server" containerID="cri-o://37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b" gracePeriod=2 Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.001577 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.082173 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8flt8\" (UniqueName: \"kubernetes.io/projected/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-kube-api-access-8flt8\") pod \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.082242 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-utilities\") pod \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.082274 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-catalog-content\") pod \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\" (UID: \"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739\") " Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.084279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-utilities" (OuterVolumeSpecName: "utilities") pod "d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" (UID: "d8df2d7e-2a1d-4317-bf08-b6a4c29f6739"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.093974 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-kube-api-access-8flt8" (OuterVolumeSpecName: "kube-api-access-8flt8") pod "d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" (UID: "d8df2d7e-2a1d-4317-bf08-b6a4c29f6739"). InnerVolumeSpecName "kube-api-access-8flt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.162428 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" (UID: "d8df2d7e-2a1d-4317-bf08-b6a4c29f6739"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.185079 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8flt8\" (UniqueName: \"kubernetes.io/projected/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-kube-api-access-8flt8\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.185120 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.185130 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.479739 4824 generic.go:334] "Generic (PLEG): container finished" podID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerID="37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b" exitCode=0 Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.479786 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bjxl" event={"ID":"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739","Type":"ContainerDied","Data":"37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b"} Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.479836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bjxl" event={"ID":"d8df2d7e-2a1d-4317-bf08-b6a4c29f6739","Type":"ContainerDied","Data":"ab0c6687a1c0f82a7968e0541d7d4185373683d62c51ac333e436f955a43ccfe"} Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.479861 4824 scope.go:117] "RemoveContainer" containerID="37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.479880 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5bjxl" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.517657 4824 scope.go:117] "RemoveContainer" containerID="754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.541618 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5bjxl"] Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.556514 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5bjxl"] Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.558647 4824 scope.go:117] "RemoveContainer" containerID="9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.605261 4824 scope.go:117] "RemoveContainer" containerID="37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b" Nov 24 14:37:23 crc kubenswrapper[4824]: E1124 14:37:23.606203 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b\": container with ID starting with 37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b not found: ID does not exist" containerID="37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.606245 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b"} err="failed to get container status \"37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b\": rpc error: code = NotFound desc = could not find container \"37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b\": container with ID starting with 37609170cb9b7311308e3de1d39ff0d20d7b3f055460d15e017c04f2f9cb463b not found: ID does not exist" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.606271 4824 scope.go:117] "RemoveContainer" containerID="754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a" Nov 24 14:37:23 crc kubenswrapper[4824]: E1124 14:37:23.606670 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a\": container with ID starting with 754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a not found: ID does not exist" containerID="754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.606715 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a"} err="failed to get container status \"754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a\": rpc error: code = NotFound desc = could not find container \"754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a\": container with ID starting with 754d6bfa27b2040dae1b64ceecd747c3543cacff6a3a5405fe17b8096f195a6a not found: ID does not exist" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.606742 4824 scope.go:117] "RemoveContainer" containerID="9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a" Nov 24 14:37:23 crc kubenswrapper[4824]: E1124 14:37:23.608146 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a\": container with ID starting with 9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a not found: ID does not exist" containerID="9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a" Nov 24 14:37:23 crc kubenswrapper[4824]: I1124 14:37:23.608208 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a"} err="failed to get container status \"9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a\": rpc error: code = NotFound desc = could not find container \"9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a\": container with ID starting with 9c6df6ec71d93e248e373c1eec7a01c8cb2a3b501e34f5c095f072e28c16f53a not found: ID does not exist" Nov 24 14:37:25 crc kubenswrapper[4824]: I1124 14:37:25.022572 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" path="/var/lib/kubelet/pods/d8df2d7e-2a1d-4317-bf08-b6a4c29f6739/volumes" Nov 24 14:38:10 crc kubenswrapper[4824]: I1124 14:38:10.788075 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:38:10 crc kubenswrapper[4824]: I1124 14:38:10.788594 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:38:40 crc kubenswrapper[4824]: I1124 14:38:40.788857 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:38:40 crc kubenswrapper[4824]: I1124 14:38:40.789561 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:38:50 crc kubenswrapper[4824]: I1124 14:38:50.784891 4824 generic.go:334] "Generic (PLEG): container finished" podID="5689a9fe-8af0-47a5-947a-b191a6954d3f" containerID="4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82" exitCode=0 Nov 24 14:38:50 crc kubenswrapper[4824]: I1124 14:38:50.785201 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" event={"ID":"5689a9fe-8af0-47a5-947a-b191a6954d3f","Type":"ContainerDied","Data":"4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82"} Nov 24 14:38:50 crc kubenswrapper[4824]: I1124 14:38:50.788206 4824 scope.go:117] "RemoveContainer" containerID="4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82" Nov 24 14:38:50 crc kubenswrapper[4824]: I1124 14:38:50.880834 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9hlbw_must-gather-m4hwq_5689a9fe-8af0-47a5-947a-b191a6954d3f/gather/0.log" Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.130484 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9hlbw/must-gather-m4hwq"] Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.131160 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" podUID="5689a9fe-8af0-47a5-947a-b191a6954d3f" containerName="copy" containerID="cri-o://5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1" gracePeriod=2 Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.140435 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9hlbw/must-gather-m4hwq"] Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.554026 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9hlbw_must-gather-m4hwq_5689a9fe-8af0-47a5-947a-b191a6954d3f/copy/0.log" Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.554631 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.660698 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5689a9fe-8af0-47a5-947a-b191a6954d3f-must-gather-output\") pod \"5689a9fe-8af0-47a5-947a-b191a6954d3f\" (UID: \"5689a9fe-8af0-47a5-947a-b191a6954d3f\") " Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.660883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hgb8\" (UniqueName: \"kubernetes.io/projected/5689a9fe-8af0-47a5-947a-b191a6954d3f-kube-api-access-6hgb8\") pod \"5689a9fe-8af0-47a5-947a-b191a6954d3f\" (UID: \"5689a9fe-8af0-47a5-947a-b191a6954d3f\") " Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.835670 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5689a9fe-8af0-47a5-947a-b191a6954d3f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5689a9fe-8af0-47a5-947a-b191a6954d3f" (UID: "5689a9fe-8af0-47a5-947a-b191a6954d3f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.865262 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5689a9fe-8af0-47a5-947a-b191a6954d3f-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.908611 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9hlbw_must-gather-m4hwq_5689a9fe-8af0-47a5-947a-b191a6954d3f/copy/0.log" Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.909499 4824 generic.go:334] "Generic (PLEG): container finished" podID="5689a9fe-8af0-47a5-947a-b191a6954d3f" containerID="5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1" exitCode=143 Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.909576 4824 scope.go:117] "RemoveContainer" containerID="5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1" Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.909731 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9hlbw/must-gather-m4hwq" Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.932797 4824 scope.go:117] "RemoveContainer" containerID="4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82" Nov 24 14:39:03 crc kubenswrapper[4824]: I1124 14:39:03.990357 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5689a9fe-8af0-47a5-947a-b191a6954d3f-kube-api-access-6hgb8" (OuterVolumeSpecName: "kube-api-access-6hgb8") pod "5689a9fe-8af0-47a5-947a-b191a6954d3f" (UID: "5689a9fe-8af0-47a5-947a-b191a6954d3f"). InnerVolumeSpecName "kube-api-access-6hgb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:04 crc kubenswrapper[4824]: I1124 14:39:04.013631 4824 scope.go:117] "RemoveContainer" containerID="5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1" Nov 24 14:39:04 crc kubenswrapper[4824]: E1124 14:39:04.014214 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1\": container with ID starting with 5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1 not found: ID does not exist" containerID="5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1" Nov 24 14:39:04 crc kubenswrapper[4824]: I1124 14:39:04.014345 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1"} err="failed to get container status \"5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1\": rpc error: code = NotFound desc = could not find container \"5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1\": container with ID starting with 5f61a98357e00a5795918bd13e2b97b70522f7ac16236c52ba80538c728b0cb1 not found: ID does not exist" Nov 24 14:39:04 crc kubenswrapper[4824]: I1124 14:39:04.014442 4824 scope.go:117] "RemoveContainer" containerID="4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82" Nov 24 14:39:04 crc kubenswrapper[4824]: E1124 14:39:04.014756 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82\": container with ID starting with 4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82 not found: ID does not exist" containerID="4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82" Nov 24 14:39:04 crc kubenswrapper[4824]: I1124 14:39:04.014875 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82"} err="failed to get container status \"4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82\": rpc error: code = NotFound desc = could not find container \"4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82\": container with ID starting with 4ce9260321a6b98db6415ac6c60cbe73c60e43c9b638f9e1263f1d7a6f33dd82 not found: ID does not exist" Nov 24 14:39:04 crc kubenswrapper[4824]: I1124 14:39:04.068897 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hgb8\" (UniqueName: \"kubernetes.io/projected/5689a9fe-8af0-47a5-947a-b191a6954d3f-kube-api-access-6hgb8\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:05 crc kubenswrapper[4824]: I1124 14:39:05.020217 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5689a9fe-8af0-47a5-947a-b191a6954d3f" path="/var/lib/kubelet/pods/5689a9fe-8af0-47a5-947a-b191a6954d3f/volumes" Nov 24 14:39:10 crc kubenswrapper[4824]: I1124 14:39:10.788507 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:39:10 crc kubenswrapper[4824]: I1124 14:39:10.789198 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:39:10 crc kubenswrapper[4824]: I1124 14:39:10.789277 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" Nov 24 14:39:10 crc kubenswrapper[4824]: I1124 14:39:10.790289 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"852247ad8ec5e01521c52f7bcbd76d7dbfe197d123f160f2a086bf954fa38a02"} pod="openshift-machine-config-operator/machine-config-daemon-74f4n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:39:10 crc kubenswrapper[4824]: I1124 14:39:10.790369 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" containerID="cri-o://852247ad8ec5e01521c52f7bcbd76d7dbfe197d123f160f2a086bf954fa38a02" gracePeriod=600 Nov 24 14:39:10 crc kubenswrapper[4824]: I1124 14:39:10.992505 4824 generic.go:334] "Generic (PLEG): container finished" podID="b90f50d7-5574-472e-9532-91b491a0cc56" containerID="852247ad8ec5e01521c52f7bcbd76d7dbfe197d123f160f2a086bf954fa38a02" exitCode=0 Nov 24 14:39:10 crc kubenswrapper[4824]: I1124 14:39:10.992561 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerDied","Data":"852247ad8ec5e01521c52f7bcbd76d7dbfe197d123f160f2a086bf954fa38a02"} Nov 24 14:39:10 crc kubenswrapper[4824]: I1124 14:39:10.992868 4824 scope.go:117] "RemoveContainer" containerID="6c4fed15b5c1dcf4795cd83284f078a823715f2f4b507153ccba3d5096e37668" Nov 24 14:39:12 crc kubenswrapper[4824]: I1124 14:39:12.005044 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" event={"ID":"b90f50d7-5574-472e-9532-91b491a0cc56","Type":"ContainerStarted","Data":"3cd129311a964ac22b809e07abae245168426457cc1916775b2726be3beda11f"} Nov 24 14:39:26 crc kubenswrapper[4824]: I1124 14:39:26.018177 4824 scope.go:117] "RemoveContainer" containerID="6336ef52a9c45ebc6fc7203300e355a8e9fb95220b69e3f74176113cc6f4f290" Nov 24 14:40:26 crc kubenswrapper[4824]: I1124 14:40:26.093285 4824 scope.go:117] "RemoveContainer" containerID="cab87cc4a44c2a3730d93fec6df63c1c505c84685be08d44e947e8ea36dd6002" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.147094 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c6kgq"] Nov 24 14:40:51 crc kubenswrapper[4824]: E1124 14:40:51.148048 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerName="registry-server" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.148060 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerName="registry-server" Nov 24 14:40:51 crc kubenswrapper[4824]: E1124 14:40:51.148094 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5689a9fe-8af0-47a5-947a-b191a6954d3f" containerName="gather" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.148100 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5689a9fe-8af0-47a5-947a-b191a6954d3f" containerName="gather" Nov 24 14:40:51 crc kubenswrapper[4824]: E1124 14:40:51.148120 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerName="extract-utilities" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.148127 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerName="extract-utilities" Nov 24 14:40:51 crc kubenswrapper[4824]: E1124 14:40:51.148136 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerName="extract-content" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.148143 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerName="extract-content" Nov 24 14:40:51 crc kubenswrapper[4824]: E1124 14:40:51.148149 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5689a9fe-8af0-47a5-947a-b191a6954d3f" containerName="copy" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.148156 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5689a9fe-8af0-47a5-947a-b191a6954d3f" containerName="copy" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.148314 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5689a9fe-8af0-47a5-947a-b191a6954d3f" containerName="gather" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.148334 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5689a9fe-8af0-47a5-947a-b191a6954d3f" containerName="copy" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.148343 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8df2d7e-2a1d-4317-bf08-b6a4c29f6739" containerName="registry-server" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.150309 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.164623 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6kgq"] Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.330044 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-catalog-content\") pod \"redhat-operators-c6kgq\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.330109 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmw9k\" (UniqueName: \"kubernetes.io/projected/48f15360-e0e7-48d1-b255-10d31b093957-kube-api-access-gmw9k\") pod \"redhat-operators-c6kgq\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.330424 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-utilities\") pod \"redhat-operators-c6kgq\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.431885 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-utilities\") pod \"redhat-operators-c6kgq\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.432328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-catalog-content\") pod \"redhat-operators-c6kgq\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.432430 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmw9k\" (UniqueName: \"kubernetes.io/projected/48f15360-e0e7-48d1-b255-10d31b093957-kube-api-access-gmw9k\") pod \"redhat-operators-c6kgq\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.432790 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-catalog-content\") pod \"redhat-operators-c6kgq\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.433031 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-utilities\") pod \"redhat-operators-c6kgq\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.470025 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmw9k\" (UniqueName: \"kubernetes.io/projected/48f15360-e0e7-48d1-b255-10d31b093957-kube-api-access-gmw9k\") pod \"redhat-operators-c6kgq\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.474725 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:40:51 crc kubenswrapper[4824]: I1124 14:40:51.966507 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6kgq"] Nov 24 14:40:52 crc kubenswrapper[4824]: I1124 14:40:52.208641 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kgq" event={"ID":"48f15360-e0e7-48d1-b255-10d31b093957","Type":"ContainerStarted","Data":"577f09125c2ec21ced38e2ba793c4fe012e75077c26dc5e5ad960060fed48ced"} Nov 24 14:40:53 crc kubenswrapper[4824]: I1124 14:40:53.223384 4824 generic.go:334] "Generic (PLEG): container finished" podID="48f15360-e0e7-48d1-b255-10d31b093957" containerID="f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e" exitCode=0 Nov 24 14:40:53 crc kubenswrapper[4824]: I1124 14:40:53.224006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kgq" event={"ID":"48f15360-e0e7-48d1-b255-10d31b093957","Type":"ContainerDied","Data":"f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e"} Nov 24 14:40:53 crc kubenswrapper[4824]: I1124 14:40:53.228181 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:40:55 crc kubenswrapper[4824]: I1124 14:40:55.247546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kgq" event={"ID":"48f15360-e0e7-48d1-b255-10d31b093957","Type":"ContainerStarted","Data":"41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79"} Nov 24 14:40:57 crc kubenswrapper[4824]: I1124 14:40:57.272728 4824 generic.go:334] "Generic (PLEG): container finished" podID="48f15360-e0e7-48d1-b255-10d31b093957" containerID="41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79" exitCode=0 Nov 24 14:40:57 crc kubenswrapper[4824]: I1124 14:40:57.272841 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kgq" event={"ID":"48f15360-e0e7-48d1-b255-10d31b093957","Type":"ContainerDied","Data":"41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79"} Nov 24 14:40:58 crc kubenswrapper[4824]: I1124 14:40:58.298548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kgq" event={"ID":"48f15360-e0e7-48d1-b255-10d31b093957","Type":"ContainerStarted","Data":"e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec"} Nov 24 14:40:58 crc kubenswrapper[4824]: I1124 14:40:58.328400 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c6kgq" podStartSLOduration=2.873543166 podStartE2EDuration="7.328378114s" podCreationTimestamp="2025-11-24 14:40:51 +0000 UTC" firstStartedPulling="2025-11-24 14:40:53.227668768 +0000 UTC m=+5194.867208118" lastFinishedPulling="2025-11-24 14:40:57.682503736 +0000 UTC m=+5199.322043066" observedRunningTime="2025-11-24 14:40:58.315691828 +0000 UTC m=+5199.955231138" watchObservedRunningTime="2025-11-24 14:40:58.328378114 +0000 UTC m=+5199.967917434" Nov 24 14:41:01 crc kubenswrapper[4824]: I1124 14:41:01.476110 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:41:01 crc kubenswrapper[4824]: I1124 14:41:01.476547 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:41:02 crc kubenswrapper[4824]: I1124 14:41:02.525479 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c6kgq" podUID="48f15360-e0e7-48d1-b255-10d31b093957" containerName="registry-server" probeResult="failure" output=< Nov 24 14:41:02 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Nov 24 14:41:02 crc kubenswrapper[4824]: > Nov 24 14:41:11 crc kubenswrapper[4824]: I1124 14:41:11.524354 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:41:11 crc kubenswrapper[4824]: I1124 14:41:11.572974 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:41:11 crc kubenswrapper[4824]: I1124 14:41:11.764630 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6kgq"] Nov 24 14:41:13 crc kubenswrapper[4824]: I1124 14:41:13.510448 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c6kgq" podUID="48f15360-e0e7-48d1-b255-10d31b093957" containerName="registry-server" containerID="cri-o://e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec" gracePeriod=2 Nov 24 14:41:13 crc kubenswrapper[4824]: I1124 14:41:13.974091 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.115796 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-utilities\") pod \"48f15360-e0e7-48d1-b255-10d31b093957\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.116015 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmw9k\" (UniqueName: \"kubernetes.io/projected/48f15360-e0e7-48d1-b255-10d31b093957-kube-api-access-gmw9k\") pod \"48f15360-e0e7-48d1-b255-10d31b093957\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.116232 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-catalog-content\") pod \"48f15360-e0e7-48d1-b255-10d31b093957\" (UID: \"48f15360-e0e7-48d1-b255-10d31b093957\") " Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.117248 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-utilities" (OuterVolumeSpecName: "utilities") pod "48f15360-e0e7-48d1-b255-10d31b093957" (UID: "48f15360-e0e7-48d1-b255-10d31b093957"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.128262 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48f15360-e0e7-48d1-b255-10d31b093957-kube-api-access-gmw9k" (OuterVolumeSpecName: "kube-api-access-gmw9k") pod "48f15360-e0e7-48d1-b255-10d31b093957" (UID: "48f15360-e0e7-48d1-b255-10d31b093957"). InnerVolumeSpecName "kube-api-access-gmw9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.218459 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48f15360-e0e7-48d1-b255-10d31b093957" (UID: "48f15360-e0e7-48d1-b255-10d31b093957"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.218985 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.219018 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f15360-e0e7-48d1-b255-10d31b093957-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.219036 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmw9k\" (UniqueName: \"kubernetes.io/projected/48f15360-e0e7-48d1-b255-10d31b093957-kube-api-access-gmw9k\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.524865 4824 generic.go:334] "Generic (PLEG): container finished" podID="48f15360-e0e7-48d1-b255-10d31b093957" containerID="e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec" exitCode=0 Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.524913 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kgq" event={"ID":"48f15360-e0e7-48d1-b255-10d31b093957","Type":"ContainerDied","Data":"e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec"} Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.524942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6kgq" event={"ID":"48f15360-e0e7-48d1-b255-10d31b093957","Type":"ContainerDied","Data":"577f09125c2ec21ced38e2ba793c4fe012e75077c26dc5e5ad960060fed48ced"} Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.524943 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6kgq" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.524960 4824 scope.go:117] "RemoveContainer" containerID="e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.568987 4824 scope.go:117] "RemoveContainer" containerID="41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.587946 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6kgq"] Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.598028 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c6kgq"] Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.598483 4824 scope.go:117] "RemoveContainer" containerID="f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.652414 4824 scope.go:117] "RemoveContainer" containerID="e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec" Nov 24 14:41:14 crc kubenswrapper[4824]: E1124 14:41:14.653859 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec\": container with ID starting with e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec not found: ID does not exist" containerID="e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.653915 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec"} err="failed to get container status \"e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec\": rpc error: code = NotFound desc = could not find container \"e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec\": container with ID starting with e3ec83fa2125f066f43c623621c76eb2314f364d709348edd409ec27baed32ec not found: ID does not exist" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.653940 4824 scope.go:117] "RemoveContainer" containerID="41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79" Nov 24 14:41:14 crc kubenswrapper[4824]: E1124 14:41:14.654324 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79\": container with ID starting with 41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79 not found: ID does not exist" containerID="41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.654358 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79"} err="failed to get container status \"41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79\": rpc error: code = NotFound desc = could not find container \"41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79\": container with ID starting with 41f3e5291858eb2dc53f943fdbf21cdb043e4bb6bb7e9a0d3ed9d21385229c79 not found: ID does not exist" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.654386 4824 scope.go:117] "RemoveContainer" containerID="f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e" Nov 24 14:41:14 crc kubenswrapper[4824]: E1124 14:41:14.654642 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e\": container with ID starting with f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e not found: ID does not exist" containerID="f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e" Nov 24 14:41:14 crc kubenswrapper[4824]: I1124 14:41:14.654667 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e"} err="failed to get container status \"f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e\": rpc error: code = NotFound desc = could not find container \"f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e\": container with ID starting with f9c5a5e3795af22b9fb725855f49f9508596eff0b65daebe9f411ce052c4297e not found: ID does not exist" Nov 24 14:41:15 crc kubenswrapper[4824]: I1124 14:41:15.021936 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48f15360-e0e7-48d1-b255-10d31b093957" path="/var/lib/kubelet/pods/48f15360-e0e7-48d1-b255-10d31b093957/volumes" Nov 24 14:41:40 crc kubenswrapper[4824]: I1124 14:41:40.787732 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:41:40 crc kubenswrapper[4824]: I1124 14:41:40.788355 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:42:10 crc kubenswrapper[4824]: I1124 14:42:10.789915 4824 patch_prober.go:28] interesting pod/machine-config-daemon-74f4n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:42:10 crc kubenswrapper[4824]: I1124 14:42:10.790802 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-74f4n" podUID="b90f50d7-5574-472e-9532-91b491a0cc56" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111067334024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111067334017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111054345016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111054345015454 5ustar corecore